diff --git a/.DS_Store b/.DS_Store new file mode 100644 index 0000000000000000000000000000000000000000..f884c6896d65e1ad243b6a557dd32a69e79fd222 Binary files /dev/null and b/.DS_Store differ diff --git a/Foundation/.DS_Store b/Foundation/.DS_Store new file mode 100644 index 0000000000000000000000000000000000000000..e67e8943379ed66da888167ffc0ed0e8910c7c1c Binary files /dev/null and b/Foundation/.DS_Store differ diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897252_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897252_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f28980bd599be3eadc229a10b439e537a76c0e7a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897252_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4d35571e6fc726dffdd5534b30b4d185bca8198748069352f35937bbe6f4b37 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897255_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897255_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf4abc9386ae60f2d7b373b9c77f1afa1f8a3dcd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14897255_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d72497848c26608b6687b7010620f494853b714c2edf28034d0a3b0e1696322b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14981454_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14981454_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..935172b40cc6a8098f3e96dbee1a3532b84df4aa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14981454_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d37d80dc0846c921256e3e3b257261b72be950c981c31bbc88761e093d481b5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14982091_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14982091_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c5442a08e62decadda2c47ab72f809754dfd5917 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0009_14982091_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3557d163ece1883a617d808c8a836eefe8cb53e288e6aae9d4fad4f81bad1ab4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0014_14981988_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0014_14981988_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5aaac214b5dea2c5da8d474e7fa8c9ddd6def060 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0014_14981988_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d12297042f08d44dc5893120a62df27848065cbdfe181a1e80c5c625a5de855d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c57fa656c8965fee1613d154a8ce2987db5f552 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:812f812913ae199e2e095fe3e05405ea1afd9496baad0fb7768baaf0563ef31f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..da6cc26cd06ece693924364014070630f912909a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0017_14896159_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d3bd0438cf3649e03c4ea5d749a41f86f5b142c749270c11db860466ea958c9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0092_14877227_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0092_14877227_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebc8f7fed51d8fd29a6157ed9bf32c83156c35e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0092_14877227_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23686a3e5ddb07cc7977fbe56f3d7044343a23531845fccb04c2302971716b27 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982370_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982370_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a0625c14bef1d6855569f27ff0ae57d0ce93845 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982370_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e497de9c180b2b4f55c3852d74ce9110ee44c5190f325d1f80cec4eec28cc155 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..db50de292a64063ed7943165bdcab41d2c3d5f5a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23dcc0d3df8d5f88ac5de3fe8ba7852a007fb3b1a0848407594fdadfbaf6ffed +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_017.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_017.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7cd1a9f9e93f9341526df9e6ff056cefd5d0213 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982391_017.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e17b0623c19ac7864f202f5df5c358ab4552d5cbd1cc61203bcfffa885b94f9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982403_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982403_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..b4a28f2235c0e663e12fec00d3f2cd041458c189 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982403_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e2fb69627746e1b0218c4cb5c20e7825a871fb078c1baab334ddbcb87225fce +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5559996c7ca7f651ea3b67808b5c2e54a669a548 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d974e41e2eb5fcbd3752692dd99035b96e654a3ccd775bbeba6bad441aaf57c5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..974e3b3e343c77f1ea2b3cbecffba2573fc56d00 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0138_14982408_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e958f665da4bdafa4980f6f4833bdcfa23202f2e008aa68acac6e545c9a12b5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879799_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879799_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7b3f278e7b7f9efc79bd6ae674795eb69c1e522 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879799_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a251e26ef7cbc0707f623dc3626d53c43c5debe7b7a581242653534d3f698add +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879802_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879802_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..fff2ce7db124172ae031d0dd2b9d65fe4edbff34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879802_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c5c0b4713098ca937f120ef2f36a79a2ed09608a95d312c59c5b5a4ec66d961 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879810_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879810_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..896db1560f097a4e7ee27b31f3a3a8b30cd29c45 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0151_14879810_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbd748685b40a964eaaff30ff7ab618fa9c10bd283c742f3d6dd23feac1550bf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0152_14880041_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0152_14880041_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..74f3c9d0ee736c5c7aaf8d8317b83a62fd315e4d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0152_14880041_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7704b822eab0c09e8d76592ffaca5ce218819c5edb0c7452ef55b91ec950fae +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0192_14983737_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0192_14983737_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..aa47c6dc0e0d7f74022a601f37a7e6543406804b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0192_14983737_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3f6057a9b2ea2896c2351b8db763ef7ffb9c807f783cdb32b15238a6024b304 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875447_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875447_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..521411eff15d6276564d01fc1340f86dc9df48b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875447_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdca4f260674d30bb4735845fb21ff88ec95b9718e7aba554fcfa2baf87bd2d6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..07f059c59409257969e8d89cb11205a6bfb80165 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b08bac00dfd19f0343119eafc30a3e101268da5de74b58bf5da0d36f685dc13 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..87ad2fe42358bd070a7375d2135f163fa02402ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5173a8b42c0ee9b0b03e727894b13fde68ec118b18603ebe0213672b4b038e2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..17e0e5745f18e1d09e6a32fbd491697ef6ab518b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875469_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b83f0f065176e97d20d94509b8bcda91f3b115956b7a2151e6c678f763c4a716 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..14e03314a52fb83a9308d95c5e9653afb7e9dc15 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f230551e435d33b2b504e8c883c0469cb0f22b4e317260352361b9fcb7f6cf54 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..f5c0bacb84d8e5ab9f64ec53c9ace25cc5b6a910 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0200_14875499_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a56f56e9171cbbb5ea0eb8249a953d11afbc9fa39e5e95b7099d431c548e201 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0224_14980190_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0224_14980190_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..55eabcb57d8c5ed6e6ddd6f0acd9b512af5ea037 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0224_14980190_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2b496fea4e9d1a2db61d2740cc6aa3e964b96a38e85d373293bf1b839cbf3c8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0239_14868208_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0239_14868208_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a5fa037252ffa004e22d6c23a3c8005203bdc73e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0239_14868208_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28d68f4000c52c3528b0dca781d82bdef149fc91bd1ca233499a71558b3732fb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882717_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882717_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3067e10733f0b34254f54d7ab9a4424b05cfd0e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882717_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cf1e18b74878c95a22108214dce9dd517eecd79a64abdd08561123d9a99049b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882820_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882820_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0d4b290a2cc124ca043859ea0aa7ccb27ba6ae40 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882820_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ac5cc0ce89940652d0aa6118b463c1f5ce78fa2edc3d2fef329b8d707a6abbe +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882829_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882829_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..be262f336343504fae3280cb794a5f0bf1db6645 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882829_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f7f7e04aca80bb7374d4ef73c45cd3ac22468cbd0a30e6582d924ebc4425443 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882834_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882834_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..de3cafa82cc9b3d942b0b798ec7005cef4e697a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0272_14882834_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6cd3a2e38cf6d5b6f3e8dc8142ed765d7cdf9517881470ee7df853a835e3357 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0276_14877616_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0276_14877616_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f1ea9e7edbca81128df2a1a9930968a896ecd6ed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0276_14877616_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8f061d12ecd8ddb8403ba7fb8a8693b55e2f5a0a7174fa1e1d6c7efe0f3aa6c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0293_14879014_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0293_14879014_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6dbb96a671cc114df097a4b1f94687cdce996b5b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0293_14879014_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76fd8a131f912212b2d9a81611492a339c57a276342d08bde7ffb59cca4ede9d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..00acf2da7a357e48fdfe3389efe6869eb8879100 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9aa3241c14d68de4f2eec8366001c3be97c6e97c131898442489cad22e79c558 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..37fba8c82ca33489334cbbd0d01bec46cd889aa1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd6766cd18c60c3f1160343d3e7b4069bdbf4b239d0fc7785a1bb8a8b1ca27ac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_008.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_008.wav new file mode 100644 index 0000000000000000000000000000000000000000..dd36171d912918a8c947e9304988c042ccda5768 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14870508_008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b201189e275df4118c57468460adb25a0e1b121a44132d7f15cfcd9d24ccb9f5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980079_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980079_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..42f4f94d788b808b3a74dae8b98c13f8151ff7c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980079_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c4ad7aa356e65f926a54e71ed674d9a3f8fccb3f7fe13cef1a941c638ff9ffa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980081_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980081_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a94ba2fa263fd4ec3da8a23c86244e75bd096184 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0326_14980081_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:510d791e52ed75310e694df0d56e4693a07279a6b63b51c101b7ae75541f4c28 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875707_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875707_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..b73012014c43541613fdb2db26253ebb34af8146 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875707_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c3f155726faa28cb24c2b61cc06c8f85327a64a8ba923b1e685656f85f42c2c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875725_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875725_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d9f263d26a921a47742ce2d9bc3bcbccfd24894 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875725_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a47d3fdf40de866df8f6752794c81beadd081a2a9f965549f13d0037f4d9dc1c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875732_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875732_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe020abb36679bc955e0ac77d2716864f739c2e0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0330_14875732_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:570fe7059e241f3f25756d1c57304e02759e20224fec2dd1827febf943e78953 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0344_14873381_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0344_14873381_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a8d10a7927836d6514ef362fdee827df38ac3f56 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0344_14873381_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f01030d52950bc4a206627d4f0cacaa76ff7cd8604f7565509b0ac6a980bdfac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982374_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982374_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..5b854261364dfb262df5f25ed06c18a2c4468d8e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982374_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bef07a60cc90dacc6800fa73318624b71c5b08c2fb4b6d2aa9923ec37e4d0933 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982382_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982382_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7ee6573725d1479c1526cbdaa99f69224f83f53f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0365_14982382_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15914b7cf25351979479ffce1e4e18ecff2a71c79ca9617b889125026591eb0f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0380_14869637_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0380_14869637_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..4e4d549e0f614b5aa5b0132c30d80e5cc289a075 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0380_14869637_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23dec20418d7930f14b44d4f005c0d31b578fb90f35d98b5e3a26db6a4a206f6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14867711_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14867711_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..39360cbd8594a6cf4db81be45c127c7ef959ab56 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14867711_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db2e92e21572355b591d5a2bd85650c750561c64d24781b97209ccf48da47bfb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14869906_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14869906_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..486e8854ec1799499af5ad91665566c22bd5bbad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0408_14869906_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1877295c76c68dbe82680b78a10cc132d5d091fc4203653f363a5428ef17b234 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0448_14879317_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0448_14879317_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..55ee2bdb51d394f21fc43e95b20db94d64cb30a5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0448_14879317_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd61871fc72e2f0cd1208ea06f85e714e115f3d3520947eee4baaad4d9d17cb2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878216_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878216_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..00f1e040c47402d6a0ce7aee5b5ff2a88251cee4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878216_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c115b88df244ac6f735caffd5fefcf5c880b764c5f8b31be2d81e8cca980983 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878239_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878239_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..9f55ae96d375cac774ca63cd0ca1ba8035b75938 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878239_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:387fb8307d921b5cad79aa362c0a9871f33a3d7f98b3a9c47be585a356e40e56 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878241_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878241_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..24b01b47e20eb1c7ef0a7e1c10b941f5c0489057 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0530_14878241_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eff632837617070e507afe556b2f6e68a557ea71862b2dc015036e01ddd1ce7c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0589_14871577_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0589_14871577_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6c9bbff377d5d4a749fc2fd2cdb88e5e62d5588 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0589_14871577_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd48123c22d6422db211773d0a5fe9c29a5dda497377db917b598acdfe5daecd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0595_14894180_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0595_14894180_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..259d19ae67766178dfd79b0105fbeb0d5ef7618b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0595_14894180_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3009aff70ae535b0006475e71dd71abcb7809605fffccbcc462a5f976ced31ad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14892133_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14892133_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..6fde7615ff1eb77e1cc44ada85634604b822c161 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14892133_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:291bd0cc5aae0d44f9cbe68f47b6721596024d695bcdc4a3da31221668ab7cc4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14895179_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14895179_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..7b5acf8cf4a94d178ca2866047b51b2bb562729a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0618_14895179_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e810eb91aa596672662976f1229e8ae20e61854ada8542ab0437b16be569509f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979553_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979553_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca1346d63755813f3a7d3b7f4a5315cdf1faabdc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979553_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88d898e864f22d278b7fa078a176cc2de2ddd7eee39f43d14490ee7c364c7178 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979554_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979554_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..318a20cf955b8e5b36f55b43b50301498fad6bb9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14979554_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:157b841576dd6dad8fdbb1906e1ea67e861449b74a820dea40db292b251964eb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14980872_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14980872_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cbfc8e087a792617c7a0b3e589844edc74b81986 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0630_14980872_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43aa72165ab5afe2cd275fbf19f03789b22fd7ebc6e62746f4d35ffb59b404f4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0652_14880602_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0652_14880602_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6adfe644756edbe07a8e5580ab5a9c1869e2a535 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0652_14880602_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdf4d39e339724b83fccaed3183d54d73474d4c0d73cde068850d01489872087 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869727_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869727_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..db98078c6800c1d476f273502d337b5cce38a06d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869727_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94d4153762459f1998a53c9374ddb3158c409490c18fecf68eadaab9b0d98835 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869731_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869731_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..ed6196665ab3196ae67acbf867489700e292bf4d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0653_14869731_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82ecde0214ba2d5c51daea8fea750ae99a6f8a7798ca490cfa283ded1f038e52 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..60e7aa697198b195999e73a058ed2e46b5a35be2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d9cefcff9b17707d16bf162da1fd6958ee9b7ab99896f5e403d5df087bb5626 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e68b076403742559d1bd204f00bc008f5233a161 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1ff4a98bb2203298a9479f879723afc007a06ae92ba105e75674ea07888bd28 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..bcab04ccf336d8b7ee0e3c8dd1066859863de335 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0658_14869138_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6746d977c852c2474551f4a672408bedd6220923c452ac4c725cd30bc12eda40 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14869515_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14869515_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..60889a42396033154834a173e6e153d04e576532 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14869515_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cae1a527a65f0e2f6cc325833a9578a96babaeec53610527346af53c26c3fe0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14979949_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14979949_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8330510462e212ffbc94b22cd866eab145f9a36f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0675_14979949_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64df54b394a925ac20553feeb84e59311f09fb86565694167a2a803ec44058da +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9ff00114120bffd03af3656fa2720e19d490926c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbbe00b281765facf37fd933f55c8cd5ef7c4a20bb2487c58bcda40be6dc77e4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..32549065c431280b59d9d455188e8a178dd77c58 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f14c402066142392d9a91fab3963a7a12b8bd9b1d6c4744e2c5cf4bd37e9c36f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a92f98d9e7cbf8751ece3cb8f6533444f43957a7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884726_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f16b28ff3d7bf80d0af5254b610fe413a0ff31f8d25f2192ccd9dc493c70004 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..bff51372a626b4276744f66b37d128fd564fa6ce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ad9bc921b5b67d3a61c69f010293c8fced72231d688a831b408a8db5c5ac0ff +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..1543b837433c860f221ef48ceaa8671d5974d4fe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884773_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95e0fa7c87a1561757b1702fc60b3a1af055f4461b36d622aa4356c8f5ab07d8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..bb93d80864bd184d0b410e802221186827f0f82d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7765df9e915040ea92395552040ee3079be24a75fd126e001feaaf0cbeeb8c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..60cc51992c7c045777454995aa6449fc0c2ea261 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884826_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e35636d004b8ac46af52665b570a0c4b6a544cbd319cf5ac8a4898659808d3ac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884875_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884875_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ad7f5149bac0ed9d38085fe45950445935e9c461 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884875_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:446e6aaec11e4a68324b2e4b36c4d5e17a58e2197cd21f266b98a86cca67a4fc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884958_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884958_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..0a301642aea8a6ca532120eec487897af9bab29a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0704_14884958_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:beb8b16d9f77189c9fa924fd2b885de11d064159d2db2859f60c90f07f262182 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0760_14891487_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0760_14891487_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcb1a9c64aae02130b95a24c8f5e0fdeac77f310 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0760_14891487_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a772ef2da28994ca3b8ddb0c7f8acabb109f290c83c2fb807372a288d4f8ee6f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884551_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884551_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fec55ef8a17ae12b5b01c39ddb2b2c294995d89 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884551_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0eff3bf934efbadfe919f01aee94521aae503a56c6949d78582f0a014d21f48 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef53f55ab564c32bd2e338856f75d4b42deda2df --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34d0d4371a3e60ef3a5538410385e8b9810c3e269bb48cfef16ec2163be67be4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e6584851e25cb20dd9f1818a0146cd060c08b75d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7292d0f715d5164412ac6685d71cfc126a747a1eefaf030c25768cd8ccc8bcd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f205d0e624749b5a1051e25e31770f2add85a9af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14884557_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2db727dca9d155d49ef4fcf39ad09422766c4e5c236b562abdef4d45d5c92d86 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885404_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885404_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..98ec90742f5b4c4660259d4532bc5f679f6c3461 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885404_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f154b3f1457e4d24271a2d5450a5fb3df8217f6499db94b2678bff78df87a96d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885410_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885410_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..87f2f1951b9e183321535806c88b06642a99dbbd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0762_14885410_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ef6d73ef942beeaa78281a110842b7c025291b2c69dbc4fd0529729c382cf62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0820_14885789_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0820_14885789_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..90ca5b979efb78943c151e471c447bee1f5dc397 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Bus_user0820_14885789_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a7858d084044af6698329b71fe75fad41251d6cd414af5a499e391df5e8a0a0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0001_14981568_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0001_14981568_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..85761ddbece27eaab6c597fa8624115fe9d29a36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0001_14981568_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa8392fb6c88387165cf2ddfd68e172ba31d60240e66a56bde9ae8311bbe3de6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0017_15310943_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0017_15310943_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6796b76dffc0f3cf681e1d3102d92d793b7e9094 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0017_15310943_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41326e5d359a5beba38707d09742e93ad3ed27ecd2490635c9f1b215aa69a60c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0025_15441137_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0025_15441137_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a706fe6cd60c7a24c5a92e4d3197b3642275f982 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0025_15441137_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d89a45c6738adc8c35fd3fe449d8501f97102b2d60ca08812fb98d277e93bac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15311842_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15311842_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2975f7ee296576652f036c0dea08aed3ea60a55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15311842_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4663e0cc3d3d4d323413d9d64dbcc1d24312e6c941aeaec9671f453ea4ca6929 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441285_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441285_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0b88221aa698680fcef6429dd0cedd49ab1d05e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441285_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ffee217a87ccb3bf7bff8860cb21358851aaa5ba4fad4cb83e9cd4dc9fb1ef +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441317_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441317_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..96464c7c5aae405b4a2ab71a1cfda5528687f711 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15441317_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:738e2c0ca12caa59b762db2f56732c6a5c7294bab637039ba8c0341f4b31f744 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15517570_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15517570_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..505dc6fa30b9b90839c35cc72cfa9e44a324e16d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0027_15517570_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f56c74e5e5bfa1b286ed6c50e719276f7439ce825bb1a102eac604944ce3ea90 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0044_15441275_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0044_15441275_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..015f8fbf87dae7c140cb769a17e566498b90d1e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0044_15441275_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a3c2a706cdeff505dc196882e2e2a995396b1eada5bc46e403848dc7350b4b2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6427d268c287ea848990ce441f0d5384d46f7183 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76b08741b365106323a22d67b2774ee40cd7b30a9533fbdf34316e07fcbc48f6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0596daa36956e01a8cbb4b832b96c24b8b2af7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14977311_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b86fdce432969f5ab838c406ba0325f4694d2c68e8ba690399c78cf31d250c1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980612_008.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980612_008.wav new file mode 100644 index 0000000000000000000000000000000000000000..5068551bd1c74dd93d1e783233d9074adb538bfb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980612_008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25444bbeefa2d6477e48e494cd028090f41d38da62d1d20fdc33ca904ee816cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980614_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980614_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..114340ba74a0cc40fae25d3f0f73ec6eeb89ee1c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0057_14980614_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27c511ba26a513cc04dff4aeb5eb0f4bc6a6270225472ec3de7058c61c09b83e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0066_14986521_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0066_14986521_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..75802bc653bbe9e1a6576fd07d1924744b697df0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0066_14986521_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03258cad7a0a0531ba36a114ae63f25a08251d2ce0c97005ca3b95ebe1cf931a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15396574_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15396574_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dab22201ea3c1a595dedcc691ce54cca9bc44f53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15396574_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7107158c457ce8ab61cae371e8591c42b45c6b9279025b934f483e86a5e2b918 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15519053_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15519053_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..65366964da3cd41feb206650c51c18af2dc6713f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0074_15519053_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7bbabe337f6325a8c94ee4b2dae4c6427e2128dba3e106ab703cea8f3907861 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0077_15441245_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0077_15441245_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f9d20dc528a9a48e94164f3ae0d5c83e3b6a1413 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0077_15441245_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96b7609367f53402f481f9160ebc45b9cdd029fac9cd787595c9e51d6499be18 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0089_15512146_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0089_15512146_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..be5125208247cfd5e464cf0023f6f5e458ab7456 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0089_15512146_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89e345d6580930657fa5b2f4c0293610e410f16043adb9203c08d38ca912ac3f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0105_14882799_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0105_14882799_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..aeadf40e5e781043c9ea46872c8177291e2af70a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0105_14882799_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:608c9d6c58530a8f2ef56df9c93388b68c4840de8d16d205db0a2b2de3254407 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987435_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987435_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0ebb2bb8f2ef61e8b72d5d60cbe4f648e63cdf7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987435_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9df1c3be657bbc163e7db9c30c2d7eaadd2c70d75df850e25fc051247e56d3e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987436_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987436_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c898d0a5a1cbe0a54bf2d2973e89b6541b8639a2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0159_14987436_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8075eba0cec773b3035f92fbd3606d32e98dbc435ef751d8ea1db90264e92e58 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981779_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981779_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..94c507678fb401be49013681ffe8988b97300656 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981779_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:136fc618df473acf9d52be0e7efd458dc0f5a741c58e134bb181b7867d5dbdae +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981794_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981794_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..df290b3f0e6eb47e9a264ec5ed567684415a802e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0216_14981794_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ceee3d5870935c5caa2b3e2ef07bfc1efb2fc48c997070a1b0fb2389519949e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15509118_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15509118_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3baac817a543a5b67937ea85a82612e903a37d5e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15509118_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a7b17ce0ae7799d46577459907396ee01b9aaf7defe57d317438a8135f42348 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15512292_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15512292_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..7a1f933545cc4214986f7c96288e54f9c4cee7a5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0225_15512292_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:961130409b977db072d2ee07959a7735344e13b3e1a7c7f5d3dd51af63ac9d19 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0265_15519064_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0265_15519064_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7de3bfd19642a402e539db7d2f4e8ac88949992 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0265_15519064_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bf0df091e3b668525b21fe395121c98e9742df9eedbcb3cfef1834118056723 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..859160772fda571fbd9af0a65301d4f7dea5a419 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b357c2aabc9b29b8179bb268eaac62555139b39eaa015ed1a6fcaa3e92716532 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a34ba24a92030583eeb9b116021ac3a723be752a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_14977244_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:689f4c963edf584d0fdd3e0efbe909013883e7bf2d001df876d40995c537473b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_15293781_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_15293781_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a08ff959836357155e6ab85635ec6ab07225c953 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0266_15293781_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ab85856a3563160b1ac2d85735d74b587b09c3fe0d23f5fa3c73661bccad867 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0307_15518749_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0307_15518749_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b6a8d6bf77f31c11f66937b4bef18b1ba1bc0f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0307_15518749_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0d898836b8992f58a93c7cd65c091a04aa889f1d787efa75717b902112c0d5e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0346_14978601_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0346_14978601_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..36fd099869edbef86052e5a1534a167cff2f2e15 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0346_14978601_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b2f90dc80eaaffd17006a3689efe379e6b8e939b574b3e85f7f908187b5fd67 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0378_14895776_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0378_14895776_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3cc6bdb2d56f1a627906e6ef2575b55376d6aa05 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0378_14895776_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:037cbe948f7215d43e0f8857c7144a1271b82936ef148b974b84fb4c302a6a66 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0416_15519127_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0416_15519127_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e9d4f40f32aeb0a589dd6b4c429027e8cab302dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0416_15519127_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6af95b1c1f1c1bb141351dc11a124e0bf504dba2e6438650d536b4d2dd703a08 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311922_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311922_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1cf9c5096c092cccc5e601e1d3f8091eba60f7e6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311922_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56f636342abf362e1613b34ab7aa9591958702b51c9630c41b5c05d60ca938a1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311925_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311925_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cad3f606d40a324d51d68bcc89324fde3a1b92b6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311925_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08634f9b660286d886ac1a766a6cda933b8eb35e41d64fb283b100b5fed46e4c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311928_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311928_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ed9b56b601b4064bff722912d4c38d7bd5e5f89d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311928_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a2f93eef50442e0a9c3c33cb75488e6de1107ea694b05695d3557fcc95c1a5b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311929_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311929_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ae69e46234460982028eb17ca2d797ee2db3f47d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0433_15311929_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fce90c080b8f080e9cd69fe60939a6c9e3e103dd33846b307ffc9cf9f4176989 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0444_15053221_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0444_15053221_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb66923364bb30ce1fc4a7b784dcc8b4b31211d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0444_15053221_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f725cbed50df11842d63c0ebf5b80fcd1dc99c77a5716bbfc4a791bd8e9e35b0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519441_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519441_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a8f249e7cca79afcc2ced29e3aa63ad12a5626e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519441_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad4e59a7870b70266583a975cd740187a1ad0e7f9ce761dbfa0d500da5c303a2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519442_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519442_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ecba28ea6c887e0c06a15a9b4f766eccdbd51006 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0460_15519442_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b376a57c6915611af4bc3f66fa7424c6647763145d6df591cae9853951a440 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0500_14987716_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0500_14987716_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1ea98c6ddde29d313f0cb0c18c4020aafa812bc9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0500_14987716_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc4e4f56423dc48771485e0f176622922b32932cb2f1860b34db0c192b530e07 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0511_15113515_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0511_15113515_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a7755973ac133d829c73e465af51af602155b2e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0511_15113515_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de7b7017944d2be650ff0cf5c54414c65d573e8d287818ce152e942f7ca94fab +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0544_14985158_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0544_14985158_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..6a5a04f433ebee5ea3fb301c88f053b650f6d851 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0544_14985158_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeb10676e34b6db98e122a0c544102eb345387c638519c784598650d9d3d41ca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359497_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359497_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b77d42c8998fa7d32d585a925fd3f3eab0bb122d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359497_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c88255119c0d10ca0da7a18a181f882ece4aaa066c2aca6983f9d909082b54c1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359500_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359500_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3dc983ce2e63db2aa7522b7e7596efb6f4238be5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0570_15359500_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06a6f2443f0d6b152b990658d0a41f221ee8d9a65960fbc9d74e9eac1cd8efa4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14899748_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14899748_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..61acfc653e28a3f1ad14e40bf55bf8d9b5e6e9ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14899748_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa31f27e5654ca14a8793aa8b97091364abb7f5c5921029100ad1533bf898c09 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14983118_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14983118_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5bb34d03802529ac1eaba27c35a82d44e6abaf17 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0581_14983118_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ff1c0b60758ad326f155eebb0c5712dfad5392f1f90628d08c5158657e736cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0582_15053531_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0582_15053531_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..152b61d608db091912ad681acc2beca05f42ae38 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0582_15053531_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4e62c6ae325fb4fb18910afc6333e967dedd864018aae3f476a8d29702d7ba3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0596_15311214_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0596_15311214_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..5477275d3c44f56979f7b25676b5d77ec6adea8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0596_15311214_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c767682d6e36b6f98fba4fe8078cf2ac50713d5bf2af4e96f95753f67e18accf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0597_14884544_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0597_14884544_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..beedc26068e580beb213acccbee5b9067a658879 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0597_14884544_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70beefeb631c4f0f0f108de7c37b9a393e72e6169eb331a38bac872642d8167e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518986_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518986_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..296cb04dd78027d64134edb3e157269f2d06dfdf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518986_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:248bf6a62afd2bc8bd8ec5e58304e27a09aabe824ac06239d08cce93609b2e6a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518991_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518991_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..564d72c87fc20100c1ee7c09b99d20437e4f1144 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0598_15518991_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:811e16fae5ce033abd9143c20e60b3bd629365db7108456b3fb712c85ad0b404 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888282_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888282_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..14aff79ee4047d27b39174f568625e3fa3ef7db0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888282_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50040a22560078f9d0d080fab9a12213cc209e6bc0a46fefb9ad71eee1da9f9f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888302_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888302_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..cbc6236664fadac36e664d442c20513d445f2345 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0609_14888302_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a25febcd0d2ae3fa570dc263faf1b5690961e41c719ec31074c4b1abbf56a02d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894692_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894692_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..83b52e78f7dd82cc435669e4970ece00c4d57c34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894692_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e27ea48ddf8b1227cc2763e8df312b491203c97aa4f03e1fdd2e24decc790c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894696_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894696_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..00e86d4bb3e82ad7ec9ea1aa2efdf2d6314c3018 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894696_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d9366b99d71254a761c0c99181c0ea08223d3b7bc08baeb9c95a9e1757d8bd4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894702_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894702_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c25c32bb0482207a50b947258a321594debf99fd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894702_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bccf3431269556eaf0429c34d01f046487d9ee0bba34daf701c66bf13fd9a239 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894707_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894707_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..66d599490511ae39129485d9d9907f8a8ca611e8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0620_14894707_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b0eaff02af41c67dfb3eb1d99302995dd98218f1cc89506bf43f83d9aa54e68 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311285_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311285_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..afa4e91e564e0d851e570ce8820d85eccde7e535 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311285_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1d62dc71fb9aecfef801a9b7832c95d5778050a90a7bb189594d78f7a575a9b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311290_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311290_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..dee070fc3e7b43773028964b2cc75c453c440612 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0626_15311290_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95ef5489234366396bf86edfe119f43956c0a618d0940640d634c2d53a71bb4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15304616_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15304616_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2e1f3d0b44e8047184297c9e21289c7822ebf20 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15304616_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f5587e6bdbbd975156e73913f58c675d69869ae3250712ded577b0cc5a08a05 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15518735_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15518735_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..26e7e40cc9c10096940d98e83e16ec39b7d40d23 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0650_15518735_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52f7883f7b287d5d5afea5923f50d30cec46da640d2700a7c7c96b81ec0cc0e3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0651_15508523_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0651_15508523_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e1875016318f4586fec1c51b8c95f1fd4f7e5b90 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0651_15508523_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f27bf0ddc323784f72ff9d077a73e47d57221fb7407c366eee126e0a739efb7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518852_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518852_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b2768d905b9095199d1d9d494edcd1e12e42d605 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518852_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffcbb7e2b51da4eb604aff9b052d19ec157af49e066f51d955cdba46b8f31d55 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518853_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518853_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..32ccefd03d61d00ecc899ecaaabab01323851ee8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518853_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f80a64fe503cdce035bb93acbae3acda3630adca41ffb0928ed440f5ce52b5e4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..d4848dcc67ebc2fa2ef114d6f210ceaf50e8ed9c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1654bd1e16bcdd958e726c5acc127bbb61a14307d445f52354cd811bd4bab1bd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ca22aa2cef237bf1061184743954e4641b9a400 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0718_15518858_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b1a23ddd62c763e69be714f046b52676bc3ba3b2e2aaebe6445aabd7e609711 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..aac6b0e21a5716114940bde4f2d7390682faaecd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35dd18f91a8defae484301379062c1b801f007fb7161fe8ec48f4fb9ddb659cb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..334be03f7a0fab6f44b5cfbdd6b382845aeb0f75 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0795_14987030_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e54bb23689dd496c95216bd7efb863a2826d37c2776b1ada4db036108882c83a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..64865ac9916e7d07c393320f115b687c3e7800aa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f92f5f2d7fecf5f30adadfecd6af6950558938f600a294271761310bee74773d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..014dab38268cbbb9bcce44c172c35466fa6bbf92 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0802_14892056_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54926732ff7b861faee627055300f1e184b7ba7bf2214e4c8ce2368ab637fd25 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0808_15519080_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0808_15519080_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..106ad7c8ce38ad99f7c61e9367df9046b89828c7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Cafe_user0808_15519080_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47d097c46b2b2f46affc89494e5298057416fa3ef6f3a7de63673fe44272d8ce +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876892_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876892_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..aa0f7cb3d0725be8421a018cd88a0a3251518a3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876892_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c435af4c51a4e0b3676a009392373ab1d859e5bd8561679a15274a60d24683b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876903_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876903_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..29ade398d17474a3e1123282e65d52085ce8f4d6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876903_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be9493d49eb55d91c38f171553113ed8474e6427ab93685693494f02eeb9a289 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876917_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876917_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..432a2cc5e8418dc997549f77fcb8cefef9acc50d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876917_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a146822348e704a307d99fb7f0f1b7bffd4e590881e59a7a6dd286ae00b171e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876921_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876921_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4a675aabfc6ca2e6f43c4af594ca3967cdbcc838 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876921_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ccdac0a7d5740fa289e58a7f9bc85c613d3fca106d4d912e2748ecfc857536c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876927_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876927_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f3e5100309e1c947bcccf25e569c8091219e9179 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0001_14876927_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bbda2b115332d54533fd565a5adda40fa3e24aae387172bfa23220ffe7bcbe0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855563_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855563_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebbc5568ae7d5223189c1113d26eaf59374296f1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855563_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bf22c0b5c125d173262dee5b73d8d83e8d6eb1eed77a8cf9bfade586730d6cf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855582_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855582_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..39a961e1daeea6e168c816b4dc98422ea9f13fb4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0007_14855582_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01540ce7f9a9df93a611e2e1ab5af743de739299dfd41e0aaa07aae9e112b9c9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0015_14853405_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0015_14853405_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..d32d2e9a715791e068342d3f172fd9927eaee55e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0015_14853405_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa1df21aa8b5836b5282ac8b2afa597b09b2dcc624623ae791c990581433d4cf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0047_14861777_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0047_14861777_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..43cc0c12f3d955939a7afbb29446cb0f31656f34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0047_14861777_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90c72c503dc3c3d7d9e1faeadeaae8874f63f0369f0be62bb29b4e5cd7ca5ad7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0049_14880117_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0049_14880117_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..b6b558f3b478f0566e871c08fe53ecf75a6868fe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0049_14880117_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c6bba148031e52ce6ef5e7ab6b1314c56c7e2966859943a2495ee8534cd8569 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..98923c38a86b9a4604e3424393a06ed19d443a1a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:070f0b9cf4f0cf7555300ea7c0efa1baec163b713670d58ff51d9626095934ee +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..4105f881c98278c73e50fd5e25a63666b863be6b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874961_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99ae6939920bfe31093eee62646dc0df25cb06e34d738752be6b253451ee8ded +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874978_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874978_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a1eed73d04c8ffac3ff6ebb5d8a108ea482b4c5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0056_14874978_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9652dfa3b71fe7e1d5c363f156cf73170ce584167fb2a5782e1e5cede8201f2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..f7dd6e0dd3a3c42b0d300eec8b75479450085d09 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e10e3b4dcb51d866d543e1e5bfa472997fe7f8f49507f2ec016adc242398b048 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..66035f6a7cb0104e9d82c63ee2a3e5e25aef8df6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855308_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:303fabde2e8fef465af04867a885f4e67e409aa58de15084447a3db4978605ed +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855337_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855337_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..36f1f980899b2818db46020582179fae320ca5f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0070_14855337_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:513e1f5985f75def32185253f031ac6a4aaa660d9ad5b8a6a3cec1acb925d3fa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852509_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852509_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..fbda659bdf9afa605827c31ad227dc514eb25426 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852509_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7d1ebc9094a2f7cde22d5c38bb5f440468da41e42ea67ea87d5ea9837e9b107 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852536_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852536_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..8fc876943fb819eb7dd3511cede8f5dbcbdd2ec4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0075_14852536_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee233a409375e990a6a5ed0077a59aeabb213066bcf4f5da2f977f5e2a11102f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..599f780f14d6839a969bff0afa9c5344daa8239e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1b47b6d95a73c4c6e5c633503eae7ccb48238e0507597b0169c57f92fbd33da +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e0b499e5bd45fe9934ed10056071e0664fb59d1b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0095_14881310_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5d0bf26058dd20929fbc2415fb2f8872ab6360d9870fc6fa23d1930cf41ec01 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0126_14848815_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0126_14848815_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..54a5dd1a3fe33a0c426123fd602310358191ff65 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0126_14848815_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc2b99bd217b89b90115df3948d8e7db8cca9a49a4f9d935f9dc6297ab9f9a21 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0129_14869765_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0129_14869765_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..5335ceb655667f207c92c862531e972509cfdbe7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0129_14869765_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88ab1bd9baaf46c464b179984cef64c2b4e91a69f9ce392dc13fb5667354c321 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0159_14867880_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0159_14867880_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..119e71ebeac6e19eb0b35f3bad2843b6e5e31c94 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0159_14867880_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42230374cb5e7e13c992183b0168a2a44cc08ca001220bf864deb62b9d5dbdfb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..d3ba20fb6dd5b193c8aa28a0358092920493538b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99f0a5c44933735c9313e98481c770f4e47c2ee157721b9b4da9d6c0d58a5b3e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c38f3c863d089ef2206119e23e2767faeeaf9a4c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14846838_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ecd56d94913af7884b7b3f86a24607e72fa153ff8e213c1c33c280ab8e7300 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14853931_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14853931_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8539885ab5806b6ecaef55bd6e0a886c3b5be22b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0177_14853931_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d20acea64e3dfb402ad090c6bfc64c701cc64f8eedf2ab67d2c1e491590ddeff +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0180_14861260_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0180_14861260_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7c43b808fb356c917a73a320e2c3be15b8bad79b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0180_14861260_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3104806ffeff10233ee1536886c214fbd6c1e7a09b5132b62f55c52c333f4bfd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0201_14883783_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0201_14883783_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..1622789769d463e029cdaae99048e242dea10d4c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0201_14883783_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba56313bae0b0719940ac875e56ee3bc7113418a8a6840f211bc428b9695ee74 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877839_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877839_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d22b17a73fbc6c04f00aae4be0149be37c3f556 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877839_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba178b7e9fdcb7fe3d05b22d1cec90da4c10eb49f984b6efc0fdf6e06aefa4ca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..875b215449d17e4ff540e9fe15438c34c2610258 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af661841511531715e5aa6c822cbbb7b97bc94c81c0608ae4b5f8d8dde9397f5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b4c86108c513044379fd59dccf942ff702633ad4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0213_14877842_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c4ad1f4f9fa41836c5ae15c9f144f86ca5c4900837978a483d8b9ad4c13d1e9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876410_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876410_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..af0255055566107cec848bdec7afdc4b94784ac8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876410_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5722b1b879aded23e71413dc3d3e38ce0894718daa596d6596e4f24a0ac4f89e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876413_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876413_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..855874925045ffe1c22c8c982e38d3b923a275d9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876413_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68704c33fd3133d2482f8b94a8a54f7f043d835882399ec6c8dbf0ad74599f43 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876419_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876419_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d33f858b57bbc782d2a3317c4eb4f129ed37592 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876419_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebd5e8f8f12619416b730bde13dadecd99ee9927614084a250ee6c76626dce06 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876427_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876427_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e66db4d205498fcb301eb95d487fab6b2a52092f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0220_14876427_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6fba6438d47fc23e14e49641a9cbc650fddfc7e745eca18b8fe99ff1b49d416 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855315_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855315_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..6bd2ac933c0781a3a0bbcb5e119545161c063075 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855315_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b6f886f033e91ea943d59d9fbeb3566b1feea139f2d73c3262f3e26c0da32a1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855440_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855440_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..1dfc66b74cc32c228ccebe8cd40cd5a44072f447 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0285_14855440_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21c3d08fa51a54350843a088f92411aa236e3bd6da07f370b0f928bc58fe44ea +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881126_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881126_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..cddbd000a1096e0cef87ebed3d6fdf3e0556ebce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881126_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cee010ec5f993291e5f9127d249c3b1efdb09792881da595ee66b135d8f5f636 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881130_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881130_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..d411b60eb03f930aab519ef96ce557ee5329c023 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0348_14881130_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de08096150ec0a8c7d7555bbad2df082b08a5044cb5485edb48de88c7662c970 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f54ae610f00878c1ba9de596c3ac951aa818e3d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a054108d00b66940f9e963a8e22e3aed4499550d25a894c09f830b82795e2d9c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..270c6ee25227d0eec482b0ec61fea4f0fc865d50 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0369_14868282_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a4941a5b53e6d3c156b42decbb87d7040ec26e34a112a0ef75745527ab69a50 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861198_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861198_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..012e7b8c8fb95b67c1cb8240b0f9935fca9eea08 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861198_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79cff32573c0ff71a1ff1577a8c4ede727497541dd7ecd0ee5d00dbc55fd6851 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861203_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861203_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..3d02e06a9bc3b5e12820a332703771cc36c18eaf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0423_14861203_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f45da3ce44ebf1c51f1b1b73b32d740b0370a496011180263dfefa33aa9161bb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882534_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882534_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..8a118492bfa26a33b749e736dafaacd7126fc09d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882534_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad39931b71d50e122f001463bda50a22242081dc5d05302b3fe0f030e272fce9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882539_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882539_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bb7b9a439b931c82982963fdedeb42010c1ce6bd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882539_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7eb66f523298afb0ef1c2f29f20fea3dec58fcb0f57b62c999662d23c83de34f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882554_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882554_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b30ea0186275391332ca32178aa35dfb3da00c9f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882554_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5fcf708d735480c77da3bd7dad36c9314021a1add68a5b6726a7adeed516945 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882560_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882560_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6817fb7d6d927dcff165baaeb77c2447d182e8fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882560_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9be1c610bd53e761c6c82e9608537118e835b4f642348f03c1453d3be96cff92 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882571_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882571_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c83535a1eebf45c0dcf7c64e439e7ffa9434e80b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882571_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f564398dc830aa0401da38c77c5ad5c875d3ff2410072fe54ff55fd97b84f5e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882579_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882579_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..adf292f7c18ba4626734a4db7736f7edccd6dbff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0453_14882579_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90d22fbfe8ef93a92c0f4e731fbccd039492f932b885cbbce2b2aed60e374f66 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14873121_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14873121_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..75e84cb0ef8169aaddde854fa5b06d96662fdf3c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14873121_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b76bb9abcf2c64d2c067c4ba9c26a190d4b0ec22e3c61f7304e8e4f2d76eb83f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14875937_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14875937_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..619ebb0de1c49fd18900ba84ddece71cf2af5729 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0479_14875937_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88e023a96f42bd2859a05ed5b0f6beed2dd80b180c776acabea88531b0698faf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..1250a71bb809259ec3d549fd0c01a689d47eb712 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:299cc557127e613ff9bba1e2f691797f506a6868f72b9cdce8e58c128bf61457 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d5d6d40d7bf7dcb7c4ea13d629004bba15c5fac4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0494_14853999_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d656f8cd658a14b7d6e2c05164576284a57c51ee9cc18c2b6be2d139ae7f507c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875653_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875653_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..2cd3ada55f313ee785e8c722aacddd3a2d486621 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875653_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4aacc10dbcb9a0e19afdf30012f9f311ffeb2a6398c0f9d58abc946e95d9cde +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875667_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875667_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..08ad4e13ce9bd8b0188896c766814d93f937bbe1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875667_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4048ec880912a5d930da73b1b014f5a7fa5569301062ba242cdad963f6a7413b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875677_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875677_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9eb923b7f09b5c71e68c611ccc3f5cd54316133e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0563_14875677_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:550406f0d62a1a75faf3586bf34ba2eb8fda0e73256725f0757778b0270814f9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0575_14866590_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0575_14866590_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5db2bb07c58f7fb4ea804f6d7911e27825206b0c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0575_14866590_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ddaff2dde1c37ba3c5837d149663b94aa66067ee9e8fbe1cd2ba2314adebcd9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0592_14862162_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0592_14862162_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..9bc6d83e3243deca2d6690ee45db5b494db65e48 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0592_14862162_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12d16f512aafd9b2d0a70ee0ed2b3c6b20ad5f35bb6022ad5b6421e14edf7b24 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869917_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869917_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e43a60803ba9607b0534cac226f1f81765bcfeba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869917_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18560c324d9f63d2dd2de1aa5eca9b25b4aca92b30c4fbf1dab96f502185677d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869931_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869931_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..52daf611b601503d757a2ec6bfb5e6ae68349cff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869931_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92138573d61514b9ff76b9379a11cfbe764618484efd08d916f3f4c6d0f0126b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869939_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869939_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f37f69b59308259f6d4b0a5571bbdea39cb19bbe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0605_14869939_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a06fb573da4843a0f26a3e0458046e52fcae1eeaefe2b4a13800091a3aa2560e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0656_14852905_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0656_14852905_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..db4905461653d68affdad7d6f87a8c44a1207230 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0656_14852905_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49674a42f3a47f81fa0fd9819b5555d23622632c8a0a26f2150b875871d42767 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0682_14875419_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0682_14875419_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..33de4d4882ee7d12b2c05897861d304f503119a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0682_14875419_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eb0d9ee2082ed96a3ab012420d2f6fef9d9449b994b59ed6fef86bc800be394 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853353_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853353_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..470c3576d2ac9ff6834461cb621fa67929df3ca1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853353_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:439ba6e8fbd6d98302f10e33163459016279fd80fc7fdf903fc7365690566b72 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..312f4d0aad6f75a4d3ffbb00a93f57ebc13b44b0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca7700d9f9d41b57e737c6ff8ae26c5f0f362c95ff95fba408d225962484595f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..160ab652b763bc635f2d7fd8cfe92e7b9a7a1eaa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0692_14853357_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13322dd87d5f5b94918c26d5223a12e560f967d1521c4b2d7b8b8be0968468a9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884107_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884107_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ae3afc91a3def8f476f6c35a3deef58ac274962 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884107_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0df385818fd0d319ad52cc4892386071543b9740a966b3734cb49d2491783802 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884118_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884118_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a1215fab30efb29ead28cfc2340946881e575c3e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0725_14884118_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76754bacf863480a53808b32639aa8b65a8de4b9236a13e9012bb150021db5a3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..30d17d799112752c15fcae4b452de0e39a2cfe70 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1be754a40cf260e27eabef2e87299399db2ebd7f3e0c942da1e63abde5cef15 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..265951f7688a2dcdc05c11c7a4e70d2be7c08322 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858760_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:646f5b6ffbbc894a5d662d76678c8ce5fc2ab77db8e8de9d0c449f1c74cf4349 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858787_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858787_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3351e4d1d746c01aa17217d3489d85d0ea6e5e1a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14858787_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae6fac1e7a8569ceddaefc3a1fc5dd09699b52c9d7e1c7eddd0e2d8e80df2cd8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14859353_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14859353_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..6f5c04e615f4b3bd1ccfa02a1e11264cbe4f4619 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14859353_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d775d6ebf52532de34bac945c1cbc94033e2e05b57c3191ccccccab3d2d2de47 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14862037_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14862037_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f30f96503bc14b21c07f0d4fe53a486f9e999649 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0729_14862037_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6428b8e5dabf366ed049702970a17c8201228526f898ccd7213c61d9c665b57e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0731_14867196_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0731_14867196_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a97589b4c644a41b2223155015e89cabba9e3bc6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0731_14867196_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3daf6ae32d02c521f09bd092e71deffed0220a9c42657c9c799304e27fbf093f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0735_14852112_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0735_14852112_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..ed4163b109aec2d0241450a69b7766b4895037ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0735_14852112_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30759aee9b28fd62ff660d880f0a0792fd52599f06bb259ad20219144bdac724 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..84b934c4d306ecd34bc98e4d9cc443f82c0d4476 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7ca0dc176d03ac1d3ac390dee6997727dff5589a14561b629f4b4687f1ae766 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2d8578f8e69279a4bcbb083b7b0a3517ca98a61 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0753_14867137_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca9aaad3787ac025adbc16eb120836df7338a036f9966f5c9ad95356d0389501 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0777_14867611_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0777_14867611_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..99feac92c327ffcabaa38ba35ea1461aef1150b4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0777_14867611_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8df9c564bb52b66c9d9d2fd15f35823637735402afc1fce4f083b5211a9ea952 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0797_14875819_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0797_14875819_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..b865c0d62c992ada452ce6871dd5095a04c018b5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Car_user0797_14875819_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20ba670b79d2110ec66aa8d7787648dbdf831da3751ba71ce05156c56b580b75 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0001_14876863_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0001_14876863_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e5189349d0539135f3cb29919e721495d22d0e59 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0001_14876863_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fc64edd51bd7e98b2ddde0ac5e1a293106b312165c5c01d5de2acee8870fff5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1301d0c948d150b9d96054cc02a6ee665274c64 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86c998962c195add95720459306598b03c142d5603360c83e6abeea2e12d9bcb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..57d7c14c9e3d4c4900093be39d2f73267473c056 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0015_14870152_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dece001092758395ed2409fcf9daff91111f3908d70a1c6630b7d9a2c7697b37 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0049_14880152_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0049_14880152_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..1948149b9df14114edd0d091b7c516bfa67ec8e0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0049_14880152_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1d260c616ae158b4785c0bee5db1f27c074e9b2db6bcfa2e31d999f0f999403 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..c2b9a8642fe3afc06211b26633a8d612edfec46a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b32fc5399ce2d7bc16cc29772209e681aa4a9c0b33ab290cfd6845a3ff505b2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..56b6d70a89af9905e30ccd7d84c2f4796c42b7a1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0057_14870194_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:341e55bf138cd3a0fabc829ccec58b09adb49627f8a43e6a04a74ea03965ae20 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0075_14862738_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0075_14862738_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6aff4594d3d180cf1d135eedae06bb65bf854b5b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0075_14862738_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:505919724da92c5ac89998a69d543882c00ffb8efef293ca1d8c8747a20d94cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871571_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871571_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3105b7b3d74a36871bb28fe9e2d2187a0a5c2052 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871571_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2271b597ac2d081f4a4c792effdf7d1cdb7549b77f0095efd56d4cd0940c96d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871624_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871624_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..24f1e4b018de610cccd216afc312f952feaca8d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0105_14871624_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5f511c6cf019d1fced801636a3609e3df21c16089dae8374f3d4ce17831f03c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0116_14869786_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0116_14869786_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef239e161a5f363a16458a3842f77dd0842cb578 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0116_14869786_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ed2c2488d701de8384b7332a8d87a9c8b6a6772789e6d1db5b82661a738a922 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0153_14980692_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0153_14980692_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..764b0747d321fd07a2fe3737564a5bf6e6907a51 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0153_14980692_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:598601323077f6ab3874921756e0be1f0e91980449d2d1fcc6142cca6ee7c113 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..2c092419278b3e37a48a4f3d9e5507ddbf7cd967 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e76b7b1cb1cbad76f325789eee32343ef6399ea1465da1f1da9c590b9de4d161 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..873cd2733b01ed19e59a94b316cf9e44f59e24b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14875904_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84041401cca9de9f4881211f83dbb13b5f7a0fd1729a886c07b60f5dd0536f7a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14894795_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14894795_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..7006d6dd23d83970ddaf93f04eb1feb6a94c1043 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0159_14894795_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c72637fbc0ade5c79edc1e75db3d2de6010a3c91b0666d3f5b46d2f5aa40af +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863171_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863171_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6e67bf2e7ef569e396625393a60e61eb1d4df667 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863171_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:545df7b90726488fb72e3708094927dd5850b320b9b4a2eea0e0b9bd7f809910 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863794_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863794_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ca2474e3c7397d45732263803eb1aa4243f666f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0188_14863794_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d05bf8c2953f29ef1c207a0ac7e3180c95e9662302d55ac1162db7c2d841f14 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0201_14870094_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0201_14870094_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a44bf046b813c62f6aac98e17faeddde99e96447 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0201_14870094_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97d00e0105c2a2085affc6bfa9775ee6f6fca15b490174840d64a59325d72284 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f23f63d135c6405a5b4a2fcaab1c9030a5c14fb2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aff7f12606a34b33acd647f278363142351c944a6097ec12a5c46b0e9cc07969 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..89d6959f9d4c7c9e1c6575c705bcf23be3264c08 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0216_14871414_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c180e52286b4694b577dc51fd0a5a28ef6df9da863dbf1a9b44a58d8e1dc5a4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0235_14876816_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0235_14876816_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3d3a301094c0cc2e90458faa4e9edcfa5fe052ec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0235_14876816_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:409cdef21bf20e4fe6fe855a31bd094a42ca4e41b476ac738cfb7ca78a461911 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0290_14867589_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0290_14867589_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a7ea9571e44f0096dd8dbd441dbde791e801872 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0290_14867589_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f87fdc1187a8717a78dc1ed38f61ac75ba050bdd7777bcea81cebc2b5e69a9c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..5fb3f4258bc05ae3180bd2415b27c00f458ae40b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5438f9d4a5d033d807feb6d23c2b192fb0b8bfad7f1a93dfc5b154edffae11a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6369dbcfd5337acae5099245859beec0876d01ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888094_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0791fffe65197b3f139b478f30d909148603f466aa107b4a406396646f4be204 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ca418532e8075e98869e92dda15c1ab81e5ae19 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:648f9691dff1d90ce6fcce69e82fd93e8fbb5356bb0c534a22dc4b2a008c24c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fa7e676aaafa36fb9b3fde4ca7d983a06d9ac61 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0346_14888131_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f90d2475786597943675dcb312b627cc66106dfde937a9c2ade886bfa0fe0bca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0356_14867614_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0356_14867614_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5b1e1f56482781c0db71622d88f34a71171a38fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0356_14867614_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30d2a0fefc1fb7808e3141bf1904d246772077d6dfbafcd38b098e6371b18573 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0365_14981945_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0365_14981945_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a1cebd06a78f68580c11d918e88af807b034d304 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0365_14981945_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92483571d3c2fccc4f802e65ee9712dccc2816358d2594c5fc0b86f737a6e293 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0378_14986012_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0378_14986012_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..94863c43f979baececd1ff06fddeb1b049028ddd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0378_14986012_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38703dbf81d05a59e8fc59b6a2ef839369042cac4adfc19cfe2199629b8ab2b5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b2acbeb7b86772ce7783c74345b3e21330a0e698 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:407cf9037ef6106a9234a384740dab1dec94aeeae6b9b99e8497d017eaecae1d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..94cc971c8eb0e295a81541aea5e09aa75ebe2b5e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0384_14977856_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24b844c5b7ec1634a1b8487bd8bd5f4cdd18493642c71d5eda4f2fdb3a4aa027 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0411_14884535_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0411_14884535_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3473ea1a4a6d0ff7b07addf20fa214af67434e1d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0411_14884535_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac110ec64b59fd759129fd615908d5e5967581ea6c5a6babcc3136667932222f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0453_14882407_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0453_14882407_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d44a72cfc9a1aec0a1e5ec56edc52517a565da47 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0453_14882407_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc023c4a3c203eebbd276530dae31b07d2f3fb5a942a5bb8cb5cda4717640aa6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894894_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894894_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..6ff310989d5eceb9600cd8393af238021794b851 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894894_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1c6a9d0cb4f9eecbcd514978d4faf708ae544b51b5d01eb6959322586f52adc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894905_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894905_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ae405cc1e5c6d86a3a82c4d499455106028950c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894905_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5dd5cfe5a55ddb886895a7573cbe4fc2be536f2ac3f30beea77ee311fc373d12 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894927_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894927_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4cfa2ad22e753815a464506193964977ec4a996f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894927_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f5becb63715122f29fce3526761ce2f747f99be043013f048bf8b86037c87c9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894935_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894935_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..0c43f3850a21a41a20b3b8c938bfe34b4ac4e45d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0461_14894935_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53e9e851a3809e5076e6b46f0e5089c3bbf678f314a018b752c0d3f81b239f85 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875978_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875978_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e71d74f7c966a595e8920a9b890ef13e5df0a651 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875978_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edec65ec28f820a8200c05cf970f690774108c71c58f9235f57880933d7607e5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875983_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875983_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb6b6b821105be520bde91a379fd2381063c1ec7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875983_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c75799d04a05a31b2c4aa54ce0713fbbcff3e04debaa202ae49bffd86afa752 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..667d5d3a70c0ef9a445bc5d84348955ee4807285 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4eef634dfa29d2d916a45df9b7712b9571ce09f38a68d1411d9d57f1e271577e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..8472a6bae8ff92e06bfcda37265d0ea4000b0c6c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3a155efece2b7e437728cc76467c7f5b64d21397d351458d51d215ea25bb660 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..25a30b91f4682a560fa54bafdc396abff05baf27 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0479_14875993_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c53f7adf5af3aaf2695f5bb5d6feab88ae079716d1099acd314c7de01cfe8a67 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..644d5532690eba52dd1096479af39d7e77716860 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b2d560a154662faa1ee0c790a2394d45f61889a219977743eaf9f3b8335335a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..77bcf6c51c3a27181ef34bc02af218b072ff3fbe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0500_14870561_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bc1cc8fe60d788f7b5f3661ce9d3089f6df785105079be44cedd514ba919a6a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0518_14983642_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0518_14983642_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c638a2d40c52a1e9d7df7a6110e15437e8f1601 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0518_14983642_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8928736e07ab19d28fe80c8496d738bed006edb10a18dab83cf476e77b62118e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0522_14980209_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0522_14980209_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c34d42475be2dfcd8d4e3b1285a649cefc22fe7b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0522_14980209_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6dc677fa521237bf62fcdb69b1a9d7da50ad8effb4b1fa8e8c9e49e7da233c0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879790_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879790_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b521a444d0ab454cff1c7262d63f97018fd3c4d3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879790_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a17235b354f51dc12e06fdcf95da4e6b986abc3fe86722d475172359a803393c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879918_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879918_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1dbfd479064dcf14b7f3c9d1e9de481e83c83128 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14879918_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5864e871e228b0165ea749c58bf946fc3a1bcd536a1c2873eb5c078c2de6b89c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899066_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899066_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..1eb3f6c51933b40356b0911ada53cd176740ce36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899066_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20d738cd255707fdb8973483e2226064d71559f6698b51b886d4243669bfd8e8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899833_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899833_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..92d6149045579430dd1d52ba35481edf414c25a4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0585_14899833_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07ad9bde93bec4d342c70ca1a5fc2fdac263c72476b72b388784283468ed4477 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0593_14984852_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0593_14984852_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..7f510c372802dc49b09b61835c77a5fb854bf5ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0593_14984852_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d90caa9d04b6922932e93e15edbd040590a7a83bd39e65d8c550ea5ba4bbb2fd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0595_14889726_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0595_14889726_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..02ee5c3af8a3282c2b3859be8b9f256556e0b9aa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0595_14889726_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b84b3c0836381a65e55049e4d9de45e28cb0654dc45c2b0d016c5738cd85d02 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986029_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986029_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..7e19553eece8fe2ee2ab665520a6aed62152ff20 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986029_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df7c544b63a8ef0fb65c18bb0ec755585030b423ad739a476ca1afc9d28b7f19 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986034_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986034_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..097220007854bbc71c357b3a005a9ff001cefaf2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0600_14986034_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dfd80121ddefea48a45e8f51abfc098bb6600c3d18b35f98d4b0928ab19da63 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0601_14895635_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0601_14895635_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a2d1475a23f4785b5153037bf3d8aa925d2011da --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0601_14895635_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa8f63fddd50a239c7ad4da4b63b7017d76f189c7f0ddd2386f33aa6cc9d0f7a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0605_14983876_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0605_14983876_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2499907c5e044343a833064d42643e7720bf2b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0605_14983876_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b873652e2f63de09506eac6592d9cfa96f4defba7ebd4b6044b3e1b98bf1d3f3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0653_14985863_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0653_14985863_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b98f003a00aa74c52756548073a56994b9706ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0653_14985863_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dff93bf58546d1e20a7e08c8fd453c8cef33a5a552d6e15495175d91736e0b2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0667_14867924_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0667_14867924_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2922d9545162a62f78110c4cb5c92d4243f29b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0667_14867924_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e43a22f62b71088dc89adb1a13807d0085e3e5004b5cbe8f11a73e0e4a08a39 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0668_14880209_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0668_14880209_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a26fd2f2b56cea29b3cdb7a0fdae75d6bff762f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0668_14880209_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4675e4fcceb08b4124fa9885898f42cffbeb79716fee6232f3908692c51018aa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877010_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877010_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..8cfefd12bfb5b404b391d852f10a7a367220a531 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877010_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7acceb046d0109fa338dce4f251e7e21c9985385551e83ab0506cf8d45217f8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877027_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877027_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb3a9304e0ccef5e27f276d9b57c4cf42d422c1e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0670_14877027_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3d671e4ccead1136ab0429e567ca56520dd3b6aede6178716307735b02808e8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877008_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877008_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a9bdee3cccde85ee733b0ac73bc907f6366e66fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877008_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e17da7460d3108f416acdb1be1a07d8a6f5e2d139f547609325a1a2af31e553 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877032_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877032_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..5b21e6372da2424cfefe9a1cfcb08d9ed06b5e59 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877032_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3498278037d16741897fccb8acb3c742886bd49521c464374561dfc150ee2f6e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877040_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877040_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ad9c17279fc2889c3daff3bca3fd8a6dc630525b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0699_14877040_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b11ccc4e498db14bc5a933145b13eeb650f87522ff32a74cc80529af643f77b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868496_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868496_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d9def126e043e097934db416ef34798065e0a014 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868496_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5345367628c0281fb692ca6cf9a91a661baaada7785512f91aa14ec7fbbee53 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868508_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868508_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..90af8fcaecb310e3079836e45d6258237f3d6e36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0707_14868508_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4a6c4619fbf17cb8e2d476d69ae5af899bf519012984268faed76e9fad95642 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0712_14867431_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0712_14867431_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d71085daad31265af7154710b7c868d18d4dbbbe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0712_14867431_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39a2b71c6fcca5df2aa3807b6827d728c988abe0a2d44ea858b5b76c3a8684ee +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0789_14869235_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0789_14869235_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4a3be88687e5b0677ef181e3ded4306bb0714333 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0789_14869235_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a34c5bfc99998aaa3f4aeefc0d24b4e6fa918b0b0b20ea0d6e7f33e9f462df9a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981731_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981731_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6c5449f595920bfbf02c7dda488efc2d8fe0a76d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981731_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f104b84db0a8f545709eda468abde2e8451b86b31e1c45ca0a3313761cc614e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981739_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981739_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..428293cdd2f3f2c63c8f91fd0fd75037ed90f84a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981739_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6108ec7f291006cac2864a5dbdafbbd0b010a413d816da9e72e58ce94ef1d56 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981740_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981740_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..a1deeba970e2faeaf9254d739e2bbfec6371ae42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981740_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3218ad04d7a4d5f4fd945e0d6edb737e1a284987a2503eec9ab498181d48de61 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981841_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981841_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ccad1fec6115c05ef7b85c85ec1282f6f3bb6488 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/CrowdedIndoor_user0811_14981841_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62ba234ff2175ee3c9cec556c9a8b8ff514a7df0cb9ce7f4e43334b6e189d508 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0004_15441125_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0004_15441125_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..059af533986d9c284b8d3db4543c7b07355b13e8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0004_15441125_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ded95460f1b4176732ddd766a34db036b304a62e8ee658a44f5c1b51a64297b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14984656_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14984656_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..28b69f88489a79239389a6c6f80d8ca66fe442eb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14984656_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13cd4f9e4f18de9ce683d0cdf4f3333f17bdf1afcb10f9620bddd5e52627ab24 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14987113_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14987113_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..40f06e66ca2440d4ecd0ec1e0e2b0f0ed5abc9e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0009_14987113_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf44f2f140e0b9b992fc5f76814fbcac2e24d7e99caec75235a60f445888d378 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0027_15515003_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0027_15515003_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c571ea6c6e7355d3d6edf951a24e821f7530a324 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0027_15515003_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08b5a7be861f1b8b29d37703c33c538eda6f1e377ada861accd1697d82a19d52 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0040_14875611_008.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0040_14875611_008.wav new file mode 100644 index 0000000000000000000000000000000000000000..6e466477c016f562da42fc4a9aeb6497d8ad8927 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0040_14875611_008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a24aec766ed1c175bc023097b655f60f0ef28d494113cbf53c070d6f10d43dc3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0054_14986082_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0054_14986082_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..9d6f3116d2ea5f66505f90fe680f9e510ae7e6f4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0054_14986082_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75a0b11d03705c8b0b664a983ac57601c07f14b11438932af5175d2db4b37b27 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15396891_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15396891_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7bcbda0ecfa7ee96147fdfec5057ab94ad514246 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15396891_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:294f526cbc7a7ebfa2372d23e27feb59f28ccfc125c0bbf43871b1011b224771 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15430303_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15430303_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f3a5ebc6eaf3dae63d8d72f2c4704549faea7810 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0074_15430303_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acc50ef1d4830626a3886baa2e00483aa3646b6b5e43859906c688945e36399a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113474_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113474_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..adbce42bca5d0f1253d49c3011cd6591174e11f7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113474_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ec861f5ded7e0fa41684194001d90a6f8588cf4e7c2f17216eb15f30360132a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113477_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113477_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1f4e619b81ae19f08671e5dd31dcbf4af430f26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0099_15113477_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5fa20465604e41f5f936a58fa635b6609215e0d1256586d4cb2cb6921b7d049 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2071f1c031310a77dae08cfee82738c9c50b145 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23ab7963e028c07887e3f5ceef7b81fa488294d5617821ac0927fee08fc4692b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c2963031839747f9ac485182228c416fae6ec77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0110_15519028_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1be79d1c639c9ab3c938dca24477c105f26ae81b3871373193883888dd6c513 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0122_15053445_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0122_15053445_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a150b8cc808cce19403fbece55191aa938ae83b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0122_15053445_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c350c9baaa49181f97d4ee64031b5bd56df0fb713f93dcaf23e7158883176b1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15508932_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15508932_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ccbdf99a5f5b8c5113007631bc63477cfe80878 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15508932_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22eb8f44756be1c45007514067602d75472d5d25ab9a599464988468bd7d94d3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15509086_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15509086_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..29ace3680e4ed72552fc57bf412e97c73c177888 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0132_15509086_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e52bdbdfaab36b85fcff9ea9d8a14dda8612455aec2c698c6b5559e75bdec6cb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0165_14875749_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0165_14875749_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1072cca7b03a7ad47c8dcb4400b07faba9481de2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0165_14875749_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86e5237103e44c14e10df568a6d1c08a24254552f584a3a561d5bac7d4e3e26f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0173_15518949_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0173_15518949_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c437c2b4bf3db3eac211f0a4ab62fef9b777baf1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0173_15518949_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a05398ac7ab75ef1dae15ead7823bd2a646c2392451d803186d9afc7fa13504b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..d8d64c69c7c640197aa2dc6963b63410fa8868ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a389bfcc7f83d770ec08ca39e17b3d676a389fe6cb37f5d2d0cf6b81c0341457 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..05e6e61c8b62da7bf0b866e1461ddadc942c6090 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0210_15400852_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20a23a3056ef385a7faf91e2b4f472330d5ea45316c9d10939a0febe0a81bf45 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0221_14981232_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0221_14981232_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..4a95a307bc08d2f11fa278110ae2f6f6ceb1ab29 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0221_14981232_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df80ab149703e3fbaad40dca900f3cea8603f80a20171e1df587cc59b6e7aef0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0223_15488245_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0223_15488245_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c2837e1cc70f0353fe3bbd7756424ee983ec888 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0223_15488245_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1c7e3301bb5404216f74341ea3e683cb80062855e0aceacd60ba8a569f3eaea +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14985952_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14985952_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..def7b0d5dbd0179dc5211567e31c052dd31e6d6f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14985952_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17c298dc1bb7f3e025ed5bf1bf99f16bd894a55b7c7575745c93335cb37bc2cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4c0a9d6101b754a0b8f8e6c4aa644b64a76e1f8d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5e1b8fef7528c934edd3a18f1cf7965b6cd80764cb74f0c20d86b43f8fe4f8c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..20a494d995a57ccd62b0af1fc0ff614c0acb8d05 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0239_14986841_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b85e0d48b52a8abcb0407fc540856e5730d48653e3584cc70e0c6302b9dcf8f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0250_14985686_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0250_14985686_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b1b0e248330f12771993d12328a82ba2a9b6763 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0250_14985686_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf00c10c6804f9e798416d18fecc0fbff4663cb6b1229961ba629719d80e254d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15018035_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15018035_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..0718c813c1194bd42a8bc719f965418d23a70702 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15018035_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:538b29d9d28ff3676d1f68a89fc0b7ed02c19efe56bcb233e1eac335e9c69292 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15306452_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15306452_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..14d74878a5bc82b7adc34cf15dfde2407d6aad49 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0277_15306452_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcfb99222f87290a8c9ad3d24aeac5d49eecaa2d7196deed13f5c4bfb7c4a273 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0305_15311072_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0305_15311072_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..0fe27125538b537b1d183f37ebb02fb34ccb05b6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0305_15311072_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d31013d6755ad82342849ad4a73ef6b823bc4f11cf1476156ba800002880f54 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0306_14896474_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0306_14896474_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c654d92e502afd00929d7ac7292a4f840c3256b3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0306_14896474_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b20aa2e3292b3c00d98f1d520ba5f1107ac871773211c9939e505d65e9c8f3e4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0317_14987346_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0317_14987346_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..34b97a3a02b423aa19380b9b087d2c3951716279 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0317_14987346_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:465d0dc96cc73d22c35b4f03231ed5877a36e3fe0d4b539a4bf18744c0341c86 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..56d4bafe68ecc9387f4b831c1dc2946fe3e171b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32da937442f7dd7a84309c26d8bb17ad01faad35497391c8da41d978ef359e3b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..34bcd0cc8b1c5416cf1c7dca65402fd68c476791 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0318_14878047_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8812eb0178a682528f9bf39929a5011bdfd276806f85c14568cb093d3e21e57d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b8b537e4e646037774c22401e4dc0f7bfaf6409 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90b7aa5caa0e6b617db05feeee4bac71a701a88fedd90cf3424feb41c89febb8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..e392f8cfc3975b7dacf2daa989a3b1dac41d1b2d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0334_15518928_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b417aaf19bc7cd7b0cc05036ed6f9d114c33556109d6a492b13a456d47ad22c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0349_14982235_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0349_14982235_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4beb9cd4fdf1e5e7d81685a6a56d92f6577b6a19 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0349_14982235_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36bd1009b09995e0ba515e98be24540a1a83a770b97400f08dac08f12e981d1e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0356_14878228_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0356_14878228_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..1a60b12197d06f05ce5c3b103f55fde6dd2d0ee4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0356_14878228_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3b8c2dbad3fd2b30b66520dc14e5f45e602e580bdc8ad83eaa4c1f0f9f36243 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870200_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870200_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c4846d631b4e13b4f3a62c55a2fbf371e7687928 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870200_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7524eddf08e98c9cbec6adbd675817926c5269aaa33f30fb2b781ed79b2cbf2b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870205_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870205_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..316bdad920c44a162a59b24661e521831df4d2e2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0386_14870205_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:379a01a56f1463535c29c5de7f645bf932fd716ef5a2b651f7b738eda15dfd4b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0397_15134570_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0397_15134570_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e8f53c681d3e1716acfeb9ecd829e300f4555e0e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0397_15134570_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5eb5762cd2103fa17fd9f587d996eb852ab05f9422922715242aa34db509918d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0398_15441173_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0398_15441173_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..a2edcdae030936957895c5e5823729c690fd2f39 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0398_15441173_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dacdea4776f1cf241d1c6d29fa373d44cd93e701ac36fad227b07265f217a829 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0433_14883061_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0433_14883061_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d5832d28eb924cdda7aaaf0b0379454fa9942d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0433_14883061_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9286d091e948d6163dc2991914a7cf5948eecf3a18923cd8584e42698b270c79 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0438_15295331_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0438_15295331_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e7be1c54085498a57939cbab9fdd618eec78f20e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0438_15295331_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:266e99a71368c1b13ef2602095c1b43b41bbf732446a0d1c8731f235e5724086 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0458_15487740_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0458_15487740_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd86497bc2bc720429654e205988da09b27751e6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0458_15487740_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25a7629dc0c1133b9ee84ed4e1a2c574acffd151f18796f1f4f0954a19af6b12 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0460_15519454_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0460_15519454_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..80b248324bc9b92443c28add9f4ee4dae5d1df75 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0460_15519454_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c77bfbc54fa041e7fb2fb1e2f752f80c9fd42fe3498fd371f2a3e60999f29843 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..626b17476e11db693c5c08b590f590d5994efcce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:302a7386f820fab988f75b166d5cb148cad50295193988c1a842a23e29d38ee1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..7902e5d9185fffb39b556a71474348b5022bd9e7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0476_14986657_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9627dbab06abf335e02e557ca4b322c5d784e68980e47349d7fec885a91135e0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0493_14911587_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0493_14911587_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..222912a9ec09081087e746db4430a3513b0ed102 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0493_14911587_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a17a1348ee5948fc43085a72558270f91bd1e47f63e36a52aba93be72ef3c3a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0503_14979170_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0503_14979170_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a9ec68e9bf2e002f11cabd7dd00bb62d3f6e853d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0503_14979170_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99c31880897a3c5f9af17388dcdb9ccfe825eff75d23e2df7b3d03bc82e44760 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0526_15119351_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0526_15119351_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..687a5e5165fef2667ab589c59d56ae52bef99824 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0526_15119351_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd2d4e00a0476c1c683772eea7dfb3d8a0d98063cdd214b22411b8a33edd4d37 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0527_14871672_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0527_14871672_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..80f27e63cc6f1015f5576475491b0a79f114cfd1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0527_14871672_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e53a252d67152c22f42b9ab034a4139b3e371200c1e9e74f501bd9b2549d3db +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0552_15518841_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0552_15518841_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..50f0b2aa14908f4ef2912e78d91bf97fd802b08c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0552_15518841_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69ae48ed98b47ae996611f787d57313aa9b7031f59d2500d310e35cd6c7657fc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b222e578abc8ff1190385c91a8659ad8e44992a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e39f2adfa30dd7ca809546edbb4fa078451ff98a0d2ca912ccc72e90c6f3f1c2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2ea834bc452bded92176fd0fedc44c2ace7a445 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0580_14884588_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1732dc6977d6c3ed6c49c26a92f3dc984098bec4e9d9f67f98e345e80c2bb261 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_14878960_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_14878960_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..67fbe2b741da02748aa3b42b509c9501521a49ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_14878960_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03c59074c3caa5fd4fbef37a4472339ffee5015d822d563b41f5cce03b2679e3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..256ef90d4664e843c6ce8d5df99e016788bc64e2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4b63553027cf5bf35ebf01e2d6a02151f819516de7aa74d49988d32ac028a91 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0c69c6663227cbf6c387c225615d8fb12e3d3f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112804_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40e1eb5c837051fd98026206e8f6dfa2f2a1cd514644bfbe0fbe085ea3e7d70e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a865a038536b9336a152249193a95bece80ee6fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3e97be3ddcedaeb3e0750dd32d54cb1604e12f35436f1611c637e1f773e1c98 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..1e677f51ef9220f957a59c198c9ca68451e1dd86 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112805_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02b3aad64de7c75c2b845d9b8f6e68021667f1159412621a3c116eccd83cd18c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112839_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112839_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..38603711388075b91b4e4a850c250fe936bc18d7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0593_15112839_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1169633f94053738269ef62a16d0cdf41da5eeb0096df830169432aa0de2295f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119384_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119384_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..865416fa5ae7ec3499e79b6ff5b39335d8a3a7c0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119384_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b247a50a4aa4aca9d1955efc9e0ef3b49a05f1c5e2c5d97f7a0de9a7cf0af85 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3523fa8fb56ee1b3f816839aac7ad46a3584393d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53ac33c55d59c028fca49ed04534ddf4dad7dd863fc6c349add8b8773b9495b7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc81cd9e38cf7cd6fe866595464aa98ea924afcd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15119386_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5091bc25350d5d1abc709acfe2374a59068f1bb68953d23386e88b474514d40 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15311989_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15311989_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..2bfc7cc2ffb9ca15d1ea26d07c813804c1a01929 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0596_15311989_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6875f3fb6a99fd5c8c2bca84465a974579b6f8eb22e335bf2e16263c1fa3cd2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053912_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053912_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..15581c4c84feafbd923204fb0742b698563c442e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053912_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:557d2ac90760a91879402d1c22825dcf329148c6c8e329cacb2c6d4567c018bc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053969_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053969_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ff02f78edecb5f8e99816b7ada7302f219ef9bf4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0651_15053969_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78e452476f8d909839488616e4d57dc9e8a4b0f432c91482ac0fe9cae75b8f8f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0657_14873264_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0657_14873264_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..2833d86407044630c9f7d39c0ff174883233c3b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0657_14873264_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:880dff1f7e2a3855cb633a6f177d0b358ea99c65e6fd2467dc29bef58aa84636 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983884_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983884_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..07b2fdb4c280baaeec9a084c49d248c348d88b2d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983884_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9ee055a50fa871943c19080cc8ecd6e31662168fd3e4deefa5ff006b530f5b8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..a98feeae5e66c814c2d8cea15a2be060c20d209f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dbdd7a160c8803e08e2a50d0f020ef170d113766e8d8cd3ead5559eb9c52d24 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..87e397683f832cb9f0ca6343e8ef9fe8ed2a63a4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0680_14983889_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a678469c1f9e745a94ad70714ef22eca9797a40f45e61af60f00172ff3243ed +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0684_14977563_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0684_14977563_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..03d417807c6b1ce7deeb6abc1b6e90e4a32a8501 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0684_14977563_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b00177b62142530e14c8e7ec3bae2184897df632ae7547c4d26e2c1e17e22bec +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0698_15505461_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0698_15505461_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b41f39a9300fa9802a095bce26ba99170c68c45f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0698_15505461_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a910b9618481f10f0d0b2495de052d75f13e37532f11520cce3d40a00cd36118 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0699_14880725_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0699_14880725_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..5f1d4264ef1822d7cd6daaca1583b94d1424c075 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0699_14880725_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7eb5307699bdbbe6b5caf0154d39d7df14061c25888d4755f1b9ab17a62c70fd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0709_15359569_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0709_15359569_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea0665631c7e4e7154c76936dc895e726f1e5dc5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0709_15359569_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76b23bc811daf96d6cfefc754ec99e7d36c8bb16ccac51eb7d22c7819df4d5eb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0801_15125894_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0801_15125894_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..cc224a9017b12684a35915335e4b92997be2e0d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0801_15125894_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e269449fd75088cb6829e9743bdc7fb3e22d905d529fb77c40f30989e426a740 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4264e50c90ad2cb76a0197c7cf01c691f3cdb521 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fcd366160c3e6f3062add8a8d27e50d46c5988be5ac6b52da89aa69c2829aab +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf70a6696e89956c23f484ccf81a1feba12e77b1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15000601_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c29ac84222d99b94177e47b3f3bf86d48ae7de20cce7972b0015d9e1407fec6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..541fac9d45eb981ea89dfe0fc55387513c485d31 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:287ce3a5c1e35a173bab690137220e2568baf61e2e72624077a415025cbcb7f1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdf7c8982d69f3ff099d2114b11bcbfea81fc35f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Elevator_user0822_15119162_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d1cbfd4d076c37356628ede569654df8aba3551f2cff3e5fd22b977984e09e7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0016_14871485_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0016_14871485_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..450c4c404dfdd60387e59ee528fbcf0307277f01 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0016_14871485_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19d002d3644b4f547dd90d6b469d7eaa7bb282040e838fb000156d2b07c54690 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0049_14834868_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0049_14834868_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e6bf31dbaf894966586ffea89bc60f1fda3fd06d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0049_14834868_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:626d52e5959f4b0cfa784f838d3e987636c6db02e4a08e7028f7c9c7870336f4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d02b2edf2456c1308a0b5f8d5bf44cc1bc1ed1f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0377be960481ec5e930c70d07941dce480ca342f617bb8743227112150df787b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..8ce2238be00b9e0fdc38bfaf3a0a50252ffb13a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862259_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:273a2d5ea85ed2306dd092d50b2b641937ec33cef20fbaa44e22b96a421be59d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862275_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862275_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a12e21f2aeb4232f6f8c22b0800fb8258adeea2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0066_14862275_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a112becf78619c22f7c422bdd70ca6591ccd9ec47ab1226ff1060328cb98eb01 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..44b45c7bbf46862d03cc5f3bb977c71c77aa3598 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:386ddf18527f6ec0119e9aa56aee4596fb6c77865755d37260f833823f700a3a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1911d55f278053386091e6dccfa222ba788170e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0087_14833019_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1705dd9e50b924d2e4380155bde3bef0e3ea1d6283b0198c0f10eb11fcb2f5d3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0107_14853120_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0107_14853120_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..387e9cf2caf64e6a1089d25b4d4e307497cd47e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0107_14853120_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81da25e1800d28a050f4ec4fdff827a3cf1f67bba5e8c1118ced75bc6acc1fdf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0108_14849313_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0108_14849313_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..40292aecb1d9ee7cbee7d9233c67a27c37116d67 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0108_14849313_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:706abbce136d4226d18b9d0954bfa6327d090cdce49c06fc6bfc178d0695859a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0117_14832425_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0117_14832425_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d1d1e4479c5775a59f94e784c18bf940e207c42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0117_14832425_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b60ad20454271c772cba14dfbb5315b102172663324ea81c327ea2640ae6cb3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0131_14841386_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0131_14841386_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e91e8dcb43ac2c27c390b30c3973af93f4874272 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0131_14841386_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25d8b4fba772ed3f3a952295b9326610d0774cd5cbbc445b846762b7d6f1bdea +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0177_14857689_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0177_14857689_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..75123b95142ad0bc8756de4c302e4853c16f7b43 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0177_14857689_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27fb3691bad919bf4d96f95c8f2bc7aef26b4887307e324a8c66ab329b2a25c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0205_14818201_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0205_14818201_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2df383bf76643e4032b9cfaea5878e5736b986b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0205_14818201_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61c2ca7aee6de9264bd53bf84f01727b6a8460569cf611499429194c6079c906 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0219_14827248_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0219_14827248_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc12d51ea98c90878fde79c72d03f1909ce289bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0219_14827248_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:273d94c25f76e971af80a2f00a13f73c7b991b8ac42dfc81f10534f3c20ecdb7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0223_14834671_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0223_14834671_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a8fb67e161a449db186b3524f2145f8c368c846 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0223_14834671_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13d98dc8026f2d25cf5d957fec39f05b2f4a0867fbc246e93ca703de9a446fd4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0229_14863840_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0229_14863840_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..823d1dffbd816e2947a57e7ffa850d6a6f2d9e10 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0229_14863840_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71453f0a07cb6f7658b4db2faf4b8764c3bf77a189e87856e408df08e7cbe084 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0277_14860134_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0277_14860134_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4e5d7b48df7a276727b4a4a12211c5cd53bef8b0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0277_14860134_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:671c3be056626109facabc01d0f65bd3a820db3331d8a2594032c31f7f2e37f6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0286_14857125_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0286_14857125_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..d783d2e909fc27a2a3e26e923e71692f16b7e31b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0286_14857125_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:691005a1c94973bce5f323476fe263a72206b93108b6992446ea39d46fe99e0f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0314_14818063_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0314_14818063_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f76378f84794d6334b3cf5d3db9191969dd67ce6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0314_14818063_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31c4b389903902c99de728687096b88ffb41bebf7a5f070376ce483c7cf1ede3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850326_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850326_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..37f67366fb8d42bafc8bd80006c432d0d5f73e63 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850326_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:930fb50dd0dea99f5a29775444d4568c4bd6e3565db0e5192cef6396d8f03a0c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850331_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850331_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..18657333945f76d25aab7b152673d8a9d7f6f34d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0316_14850331_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30e97b1dff954e2290540794c9e8bec168af660446ef689e0596ccf782574505 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829905_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829905_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4caf650e0b7434dd594fd59b6eb20b4136625ee8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829905_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c38e03105e628f4ee710279ab91bd1398b54b9490ca5f24fa8ad0e48a5fc163b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829942_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829942_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..05d5220bdda9c733db9a0766e9d3c264e9c7b681 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14829942_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bac72420d6e84e488b9dbb93adc361ab32b86c4c4d05e62797c100e5ebd75243 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14830066_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14830066_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c2de642c4b67bc6ab49ea6c49424fbffc75d445f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0365_14830066_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eb455507a3155bd4d7eecd031fb6ef8b6d31217984152acbfa361927c0fcacd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0370_14853988_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0370_14853988_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..88f1d1f978a664007bd066908c2d432e0f3535ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0370_14853988_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09a20835a74debdba760865af1d16f6963cc9df3f4446efba52e7a9ebae9f710 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14817863_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14817863_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e2f4ca3bf232a22fa95df395c89fc522b45f0dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14817863_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1077c41e2ce591fb176c156eb48bcd7e7832be5010b6474aaab64f5efa42d7cb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14845646_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14845646_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..53fef2539032e4ad266f2047d3dfd70f1ee56eff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0386_14845646_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79e323e592147b4f3cd031325fe03bd3944ff9cbe61a5d958e957b3a47d6e8e7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864393_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864393_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b29ae740d348ade73d08192da1af61d90a57d358 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864393_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:375a62d571cb9123179fe15d4d39823e36d184accb5950c46206da685167d8a9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864402_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864402_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1c5e9b580745f2ca800f2ae146b817d94e19258e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864402_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3544029863c6967c2fc5e6a1a20b4f9a5e143a7f3fc1ad8778b7253ca50474a6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864421_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864421_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f01ec1f6bde2617347317660688d1e9cc69e1be7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864421_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d97f58c45f3bb3db5afaec761e2b57c9a575cf128929dda42f55d849d1636156 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864445_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864445_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdb78c21b0cee71707b02aabeb5c86a5f131819d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0404_14864445_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b0b029233302916ff54ba225d0369f7e682d5e81bb41913c348f9405c599b0f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0425_14854091_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0425_14854091_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..7f3bc6b5475b5530e302f128cee68a9e2edfcff7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0425_14854091_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2172b2a3572b22a1242cb97841cf6b4cbf71d789cc1cd526f85aa55b7e74e66b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861699_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861699_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..8834af0e5ced64460b5d81a84e6389d1f2adee0e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861699_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:907a98312063cb514da37e92196cfa695bc3780be91fbebb3e4c96d698c64295 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..51a5a5809a7717ccc0deddd630753e721432218c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9805defd3cb2543328fdce6cdc0a526f82387f18aefc4da3b103b339d3e32bd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d46e509e1bd8a9cdeb2ae278cc3c6c981273af17 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861774_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd55402920010eddd2aa73d98886e01fb1aa332ad75f141ccd6267e07dce6dd3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861792_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861792_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..13cbdc15f7765c87c20849045007848fb563ae11 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0445_14861792_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1f5cf814a4ceb8b7b6a50b8cb0c7488e0e3c50b8cbb12faa02d24f5de1142cf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831257_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831257_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..66e10cb4e8d2514b307602deb2038e5728ba077b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831257_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b2cc0d95c0943ef00397c01aa212a1c0482aabfa3c88a912c7759c3268be37f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8fd0352f08672c6900980122857fb56641f78aa0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee046482e34a2f8551f7532a00aa0c9807a1844d86e8bde1c6811f02a2b7b1ad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6c3ee4759e9abf91c0be88e9ece48fed7516d170 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0453_14831281_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8483eb681c31ba82487ae22485f08f530050bf3a83084db8e452a8cce669a2aa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0463_14862168_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0463_14862168_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..49edc5e8d7c3488e46c1e830defb0b8fe632ecc8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0463_14862168_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4f9679abd9423429a8c6ae14fa6ea8cef1ea7441938abc5b873583e91ae8182 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0500_14826615_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0500_14826615_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b90d25fa144fb78628c8a20c74b73bfd388caa5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0500_14826615_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b052468a8c587ce23b4f3a7f1efe3a203be3b14ae20ef78a68c0752c87491f9c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834760_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834760_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f3fd57236301b7423ff95c44d5b35b6c0455b4d3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834760_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66d543bdd6d6b88cd7ba111811c9da6769ac2d118020313be17ecb181c69afd8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834777_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834777_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4cb5d07288f67b5e46c4662b1a1c511824210c89 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834777_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:338552d4d5f983fd4bcde46f3d0b0802a4418c7e8abed2e994ec454258fae930 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834791_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834791_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..427fb0b8f1bd2e6ea7c5f8fd67e8941bd17450dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0585_14834791_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e6396e4690cdcb73a9afbaf43025a947168be590fae3464fe62b76cbbdefb39 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0589_14842032_014.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0589_14842032_014.wav new file mode 100644 index 0000000000000000000000000000000000000000..00712432a3dc6a68f054063c526c017decee6fc3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0589_14842032_014.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45b4c9596edab570aa54a957880d0f3df52557703b70da8ebd75a14efb2c2c23 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14860575_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14860575_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ad82fc14752e8d39e578b24d15ec9795bcafcc90 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14860575_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eb62d2af57bf8e4b454f186421a4056a7b5c3c73c2c0c816d1501a59ada1354 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14861952_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14861952_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2a34aaa0b220c05d1167d33edaca38fee98879e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0592_14861952_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:708533777aa74c48d8a7dd1f3878980b344bb93b1bcbea14ff8c4396c51379c3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0595_14838929_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0595_14838929_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b21cf0763566490e269673fb297f8fa24890ebc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0595_14838929_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82b59815fad4c09bd4c489579f45c3f75b48be8ca6cd2967a5c819c62fdfb1c8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822265_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822265_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..76e985e346b36ac7d6888767745afea6f551b80f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822265_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7c7fb2e0d0264aac103e13a3774c927c52bf69bc2e15a0477b0a0ec06138a06 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822314_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822314_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d862e2f0acec0d1b278ea5ea86198ea8474678f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0599_14822314_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98b9eebf51baaf21c2c236d02022d5426d2d2e9de74861e79f6e4e011a864a57 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0601_14852290_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0601_14852290_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..110bb86095ef5258d2d6b05cf27ec2a99f46e60f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0601_14852290_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:329fcdf334ccd41b18e51222915307d5244d627c5d63115822324026f6463a1a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0606_14845978_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0606_14845978_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..4445960350e72719c3cd2abbeca36995497809fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0606_14845978_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eb11e96831198ab237876d1d02499851d7f002708145cc6833a2ed83379217f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0646_14833246_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0646_14833246_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cad98fbbcb82b758f22b3b90b00033d410411710 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0646_14833246_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99525c4d6d53958dc9e342e9afff2bcbbec962dadfbe55254294a6071ee04187 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0653_14861242_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0653_14861242_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fa23ffbbd7ebd6ea117a645418ab9849c226e16 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0653_14861242_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b00219e05adb9b54a63f173463cd44da3ccf67fcde059474f2b603beb08116b8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..720131b8b1f913f4cee1b4e78d4b534d520bbe23 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2059ccf95a447119442a77b29dc89be7ce4434649631a08d1286a4b16dc308c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..299d131153910668c86e98932917f7aeb7b8ea36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836362_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7abfea2efdab58008ccb9b59541530be7ec80fa0892bc1c2b911cf088bf38fc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836431_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836431_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f32b531c284066a327c0812a31c6a1e17b00cd51 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0658_14836431_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ced7dac24f1b8b3ced48756f4ce11ffe273c6103f54011cb976671b7032d84f2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0667_14823379_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0667_14823379_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..074b5e39e68a63d667364e0bc50dd9009148cb8b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0667_14823379_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8a83d51f0eeb7718f061a90b5da3cfa8a8736a3f2881fbc3bbdd2057841da18 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0688_14838296_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0688_14838296_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e49d41e0cd5fc5a5ae4532d3ed0a96d13d791b44 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0688_14838296_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b3c5a47cb68b12277af89301c9e5cbb680fa687fa71a4835e6c0a4e03e568bb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..076eaaf545a3de9b2083b98af5192f715a1480b0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6705910f516305a9f441e47576c6fe518333043f14a858f6c0d4d1cf872cc86 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..6aa4d48cd1f70e13f0401ba861395a9ae1f3c6bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0760_14833035_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d35d47a13ab8736f6fe627a20755368e55b8a3a6c81fbd3012294199c23a0987 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855770_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855770_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..9933949dc01ae82c9d95b3386d0a63e87b94cb27 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855770_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e7b01756e9e01355923a3f42ee693d69f91a9ea9c99a96f4d0a62965640cb64 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855822_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855822_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cc2b66742c15635db442e7bbf755f0e64e22b740 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0772_14855822_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48fd6405fc6614ed15ef617285fb70f563f097fde409549cc435f8c07c08da91 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0805_14840461_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0805_14840461_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..193c2d489dce3c69a9baddc78f2fc4fc6d124866 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0805_14840461_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c6eb59d69eeb39e9e4bf0cc29afd59defe431a23c492f2a848264bcf41856ab +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0808_14834648_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0808_14834648_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c7b693115c4f787f0c2a26e9852b001e08d43b5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0808_14834648_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95a8291995f41b7df4807c93ca0fe5b586276fdebc1c90971b7873ad3d024bd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0811_14854026_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0811_14854026_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..ce56123142371325ca88794f609590e7d833cc09 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Kitchen_user0811_14854026_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:710525deea8a9019e2e1ed7c771db284db489ce113ccd648c1bc59d09ba09692 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0014_15016451_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0014_15016451_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..bebbf35446a4f42aa1815ae38acaf6e8192d38ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0014_15016451_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0578c20e597e84e45c9c615a7cffeadef6253c376e70d0a0e3f786f06f29ab77 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_14983321_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_14983321_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f5abc5b1a428c17b262f3305abaf2b3acf7c727d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_14983321_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd07e5c11f5846cc8e721e23599cef1c0c4125e510523b17801ce41c13ab76c5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15053773_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15053773_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..794fe0887652bf08d4da063ff928c2ba9bec8517 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15053773_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e62bc7444321bd315a48724f3d277d5af86292b2ff34e5fc78a04edf3c978f7c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15295247_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15295247_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..3fc64c44ce2ddb7e3c7c62cc0a576d40f1aade81 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0016_15295247_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef8f95fede0d5ddb5f3ebdf10dbcbb8c2b8773278c4a9c8a99b97d01080193f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018054_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018054_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..277d08a454f97c72640b0d5d84425e2039150da2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018054_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31499b75d22a126342a15fa4828314487e8561a3529772bc9ec8b741fd10e400 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018058_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018058_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..448874e7e76a71d024cfda33db4cca108018addf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018058_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9270d674df86d06d35f99f8b069bc1420f6fe0f496ca8db7637007372a6aef50 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018062_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018062_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..93801f081d389cd3a451f2dbe7ceb402def687dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018062_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f85e75e37dc0c0a32c6d6d6d3c93b2643e5b4a54714b54f98cce495d26fee61a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd1733e5eefec05835c78753e05d99bd9384aa34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:269ae90b9c2f022c6dc90ee9346c881e9f169cb79aa91851b9c776204d9cdeb2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..406cae3494b61a73271116358c199c3ef72c86d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018064_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79d37f637484321d7006a759eff8e866673d4f0e2147445d6eef4b188fd97028 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018065_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018065_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..881702f54753b2eb0af172d681643ceb49f25500 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0024_15018065_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cb03c81762562a081872d9f4a914235f5506a7350592729370b418c20b4a1cf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0066_14986531_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0066_14986531_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4e96255aa39400799d7ae940f04ece36eb6c5873 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0066_14986531_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:662941f321dbe753eba836d23d445d5ed7250f1b2851f31ed12c6d8ed5b9deb0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f62bfa681537b2115e891b375a13fed6cd384eda --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3f3270ab8f7224c78063bb43de2548f3fa9c6c1ded08fa05e5a59c3a8f8342b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4018e5d28dc5f85a1c794ff19c9936f3a3697131 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0067_15519727_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a5028ea0b2c83f41e947cd8b36f70db4fc06286f6b05fba470fc7d3df8ba881 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0074_15425576_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0074_15425576_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cdd3fc421f2f0f1ccaf6cd385bc0641817bf64ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0074_15425576_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97f1170c8341859e394e1ab24b3084d2855d500ff74524dd9a8c79f57b3ad3aa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0076_14987381_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0076_14987381_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3046b59e52fa465c67ecc629addb683426fb938d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0076_14987381_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ef0893216c6cd12f5eae223217497d792ae2c74cbb2c81e219353e0aa20cc96 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c50a22e7f26ca4baa110083ca450407c5e8e268a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d35edcb9e55025ec280a19d5bd5a4b30ec31e78c054e4614d4e6ebc03f2929d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc3c5c16f70107f254f5573c6f991d4deb8a25cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0173_15519005_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:996017734b571a76eea59d6d8781ef629a0cc4b357a39bfdeb467ed29d792df8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0208_15519476_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0208_15519476_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..56e79e05528c1316e159ff87781dfb7a7fa7e529 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0208_15519476_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e039af631bf60b6dedbdcf5304d1cee1523c610132daf13e1ac06986e285864 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0226_14982492_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0226_14982492_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..ba0599d0fdc6e19f3785b3ee74a400929e6ecbfe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0226_14982492_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cea84acb6c503ed5da0d580f0007502e9c32e22d478444122ed8d0a9951e2e16 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f57a16c4de4bb1dc6d0b5c5da0bea29dfc0af3d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f2592e71e7197e0c6c424787a7599874b2ace597c4f926bbaa318ae7a67cc74 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..0fa5ed7d21a23ab2f6c3520b337536d0977e02b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0238_15519568_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:008b2ca8775b935deb5a527bcfafdadc38e0e1318bbcaca0e146278591930ad2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0266_15119076_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0266_15119076_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b417996293fc122e731bce34b3741c68e7c00fb6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0266_15119076_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21691365fb0d0407fbb0906d7f663108cf308f9e76daf244fcbd26f7e9092d58 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053854_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053854_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..62944255b61507c9714eb8b8fd9af09779e6771c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053854_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:941981d9d2c1d084cee5776eb4aa9098f18cac632926537953e6792b18af9603 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053855_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053855_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..34823be37e649f58de540767c02bef2bb865a79e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053855_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f375c7e1602b5c148b4f4041dbaa5221f7bd5fd733c4d34082181a6bec48c41e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053856_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053856_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..41f6c41cf72e604c3ee5270b6378fdbdec458d3e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053856_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec79831428aa0518673afcfe6018b25f8e99d829a44d9795d19d37211e917fde +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..f506fe20a4e29302ec11352f0ab681f010a1b389 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c051baa7b4328f0b3e20903467b87f20b2a7b84683b1f644b77932ebc6cd504d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..63f1b9c09dba6cf753d1b7501c2045bbf5fa43f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0274_15053860_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2c6732ddc4de9b29c01c659befa51eae20c0b55563f1b106a1aa61df088616b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14978497_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14978497_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..42a2f7ea90d46bf42ebbdea6db991f6e212a57a7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14978497_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:855788c890f7b2adf0c401f699c5e3fc18e07eefc253fefbc3fa48f1023210cb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14980341_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14980341_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..16d9e82e89314cfffb4015cf89c079c09017ecba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0308_14980341_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aec5e079423384bd287b7a30c140a90ce2485405e83930a74f5feeb0fd7a1da1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_14893811_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_14893811_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..671c19d21a2482de1d3d62cbd4351cb83659b0ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_14893811_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8f27a1910e43f2c49158c037662334e67e080b4dfa2846f9fb89a282bc98d5d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_15519048_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_15519048_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0a8aa7d9f74e2c0eacca888d0fed09fb02df6051 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0314_15519048_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fde185d5f80852096542d4a4d2ef721e01a9e4419c7ca370c486138b0b7c320c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518738_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518738_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..f61e86895cd676e3e8278a87fd0438a9aad7f37c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518738_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ff21da52416ff9cc9500dcf967bc5e2b348bbf4170a8bf161bf25145954ea3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518740_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518740_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6e3683ba4080dc3ac3024bf171825f22a00513af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518740_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bcd65328be506fe409b2cb861c53d8627c5704a3595105ecdcb6fdef3de2910 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..99a17bf4e16a7c79ec42be629139fc132c60ea97 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac2e79ec607df33557b6ed293a2593d628ea077e24b8f7ada7dcd2ad18c30767 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..6669ff03832e00c8402de53e44e6b35eab37f7fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0323_15518742_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21d665d82db249f83e76360eba1b49354be5489c0f10c8f72200d055a9301ac2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0334_15519083_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0334_15519083_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..77a205a6b188cf489b9b8453a636fcb1bac8f3f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0334_15519083_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6e212764c318d3918470b51b6150a3858a0f4695bd5b78c655b9b5815f7dc7f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a74ace424e16162c8a45a4ea50ea3fd27b7474d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:750c86d30280cdde2f9e449c6726f730ec586e30c27999fe209d95e7fe762c18 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..22c078484d341345b67adef63e13a402b2af9d19 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_14982354_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:271a656bbe1053bdbfe1c252e5eaf37b76507c1b3565e818f0ee4abf17aff9ac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_15053734_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_15053734_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..3fb6f7a21bcf826be9ee097dee41df956a909474 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0352_15053734_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:866f9965127fa464acc40f273a162aa5f498f90a71e02c8f1a0e27b4ed3e6f87 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0380_15519726_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0380_15519726_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..34819e5af91c46b2aff92bc0dd3e3edfee24a299 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0380_15519726_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1cc09ed0046f9b09f0aa7fc4e0c43ca438bc7fe58af6f9f89c548690224dd2f6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0381_15208821_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0381_15208821_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c481bb8f54c7458ca8f3c348c491a1ce1cf4e6c1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0381_15208821_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:094e0aefebba2352fd4c5fea18078fb1034061e91ad8f0c528f491467167b1d2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0398_15359579_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0398_15359579_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6c66f25d91a77f0f068f0bc1ca493c2b38becce5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0398_15359579_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bcbf0bf87d2a0225d15ca78ef6f542fb0c9eced4dc7d5c35c4a88457f6b294e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0429_14977599_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0429_14977599_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2f2dbfa4e2dc5ded7ca7d96a2d9db86183185d29 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0429_14977599_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4582e7069be5c92392742865e6860be74e74d2cd358c5982f640f48833f714e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0448_14890871_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0448_14890871_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..0a36a794852cc390d890b079ce8d69df5594d95e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0448_14890871_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:230f7110b7bc3178aa43f33a2366e65ec2bb38b41c8a824895993c48d5d09a87 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0460_15519894_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0460_15519894_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..97faa093cd12395f97b6d0e1e8dc9982e12083c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0460_15519894_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3355b11ae25fb503761673de18f1db755286cf944f49effdea15003b63652b4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119288_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119288_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..79e2cacf0b897c2666239ed99d1ee26bbdaa5ad8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119288_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b1474c5a08f28ddc0de5dd5c89de68aa70c316d9013a52dcc4f024aa5f26c21 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119291_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119291_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b178fd81195928b73c3980aaac3320736e54a1e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15119291_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:900e2fd1dc6afe0094894988b609498660c608bde532c6e93f3a18b85d0bde02 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd62b012dfd5a44f11f09bf4971ce7d0c285d506 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fafb625fd75be3497cd1a6735329e34a2def151b9d8830ce3103625e79039c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a846c9a8350333881565660d2875bfc74cb2cc3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a93b9b50fa8d4af4dbbe4b69a6f9ab9e3877a7e8511be5eb9f8521061dc4b79 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b804ab46c916dd20c294017527f8581e5829ab8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0498_15520217_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dc522c692a31f060cadf3550a53a6e51a726fbed0ebcf524cb413a51ebee073 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119568_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119568_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..59d11a8669383120dbb4130d46c6ea3dbf600dc6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119568_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb1dfb30e21c442ba992550796fee2982e7a4653d556ab9181dfc76f58d2d1a8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..7bcc7b5f64b5aa5c001678d5120b66a7d62af8bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b18fbbbb4a458981012a067d09d81bcd7a915ec76fee22b15abcd5b92505162 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ba7a7fd7759bc96ecddaffb426d443fcca513bfe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15119570_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c7e62aa135e528ebfa91dd5a49f85b8d943b8c81631cca0538d6efbb3c7c2f5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15520179_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15520179_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c09111204c89d86ac9f7d74d7834dc3a246e5e8b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0526_15520179_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd6891f9505f6c4dc06dc342afafade90be705a6c2894925af753bc6a63295ad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0533_15119089_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0533_15119089_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d40bfe87f85aedc3aab8e7dfc4a888f04427a53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0533_15119089_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbece3aed87bb4399af828575a8220237fde59f80771a8cc56629a1c1114fe68 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0536_14985881_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0536_14985881_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..01c97fdf356f41f4281f782f99ca2c55bbde48fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0536_14985881_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:657fcca36eeee2804856e19c2acf97ce0693dba00239c4daaf78eef3ef3b6e4f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0556_15520028_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0556_15520028_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..d8b841137100c91240318cf5f7d26e0a20981147 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0556_15520028_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c461f6c3246c3d3117efc51255521ce652f487d1977dd2bc4422abf438fc604 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0560_14979261_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0560_14979261_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea018979db872a058e836e661f841aa0fff42d02 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0560_14979261_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:241ca0670d0ec81b76d4ed8dc1e1d8bfd1b36bda0bdaef17aea4ff61a24e776c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15119596_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15119596_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..99e98a8f0c46d41d4a53fce5967cf9e8c072ce97 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15119596_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b82aae294a65ccdbb2988ec8b2415bd1c3cb88411d5a026c024d7c995c0b0dd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440295_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440295_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a4ff8dada7039e63d820f1b820c5f2c153cd9df1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440295_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:465c939d3118e104dd5e59fea26ccbe57d1bd0577773895b64e40ef718cdcde4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c47baeabcbefcacdf29c856e0fe92dd8ad664161 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2ad9fbf917ce559c4eff5d6c97558d4f3604b3466de989a984fda7f6d03b9bc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3677fb432180a64d9d02b09d111ef76fa629b14a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0596_15440296_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fda4b7b76f2ae280c479d298211e85a29048a7f4f4803c6e1a66ef4b21ba3b6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894892_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894892_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd8e6eb44f382600b9f370e653fa518b612fa4d9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894892_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b18f454ccc36b1c3749dda3ed221df709c7bbc8f6f9a58fb578ecd6bccc7bca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894898_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894898_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca72f9d425fd29eac346b892d4c2279e50e1b4c1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0623_14894898_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d50a79f42e737aa6676e7b540e80d2c3cfe0b1a8c071419a18a98c3f923d1c18 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0643_15412392_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0643_15412392_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f132f96315411771f3f2b23f0eb6c92091b67c6b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0643_15412392_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb90ba09e1ed550762ef3a554489c98f421c5282b062bed248cf4ed82018e936 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..9b1571e0da68feccf0e6a8e46bec1dc2974e2272 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:135211cae5fbb0745246031eabc4302e431ac0e5105670e9f342496c46b5b951 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6085df5cf3c62fa6f1e30772d531fd574dd7a230 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0651_14978862_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7f3cd5416b578db4d750bbcb224c3c26f69666531e6a4eb6b30007e7d3628b1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..677c0ba8aea288d4c1fdefe338d5a8fbfa42f20e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b65600b73fc506a9753042a1c417e7ede11700f555c98b16dd739e05e7f6e7d6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..6de665a02c2d8f511ee932f7c56a0186718f45e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235517_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c7859e8e8fc7dc359f6c4ebcfb49e29aa00812b32f19b038876bfe2d81640e6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235773_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235773_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..62e6739a75792398b1c4e8c36adf04e88d377871 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0657_15235773_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6e6825160fff212d25eae3bfd817051582f704450c074f059abc5633fdd973d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15053571_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15053571_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd9c696ab639edbb60794c35c32c8d9d04bb358f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15053571_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd9c76552285f5864080e77960ebb1c48eaab09408d4d529dd44bbea8b3d2628 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508524_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508524_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4100eb9bdd45ff0312d4ffe5bd32b8f75a69a8d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508524_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6eb697f94d336fa65f9c33e24c69f87a44b57829f6d851bcf38b88db04fbb80e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508525_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508525_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..d1214dacbb2a376cab8becb56d173e374dad6a7a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15508525_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40bc8214e62fc5b551aa6664ae65a403ab192c50e8c2919def266dcfe9cc53ed +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15518993_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15518993_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..c80dbd5d6a52006d272b2f13be663b2a6157b18c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0698_15518993_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54aa953f64aece77b2275995e4fa8a1c47bf9231f03aba10b9cc5ed03100ab5e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359575_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359575_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f74729ebd8e76ed9564214301739b0c6d57ae55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359575_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3da31e56554f2a57c40e8e0e5b736b93bfd01a3b2f00307fda1a8c5c6cf7ce9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359585_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359585_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..8fb4df62365e864191dbc18d29bfd9e71889a75e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359585_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:446658889e78fa544a56c5e8a1ed702706f852babaab2f01535e38850228aa2a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359587_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359587_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..217cecc68c7f922a9cef108981d50cd5a9ddea20 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0709_15359587_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fdf3cdd5bc138bce478af50ea9c2fc00ce9fc161be14d48542b38958e53c19c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..55abcf819c6482c000bc3220893e1940c370b42e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65097afec36b1f30824cccc90fc2f2b6ae50ae587c501ad97fdb6f2424112273 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3256e83aac8690fa62c7c2c1379bfbb11b81c2dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0741_15518846_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c491f59c49b3163ae4eded70621f864f29081c43679ed8db1e6275036972616 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..520626354841828a7a5fd3f08eb26b5d4fa6284c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7db48d3f4842cdfca70fa7439f6a06a0b3dfb8ba711c9c4de539b8c3325502fa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..196dc87a1d711ec338acb6c7116a12110a420e88 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0769_14981009_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87b35e9b9253d279246c1f112c74a797ddb138e83631ff0cb3555979b5b4b513 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0775_14983643_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0775_14983643_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..9b0b83c422ec5348df2d2f6e6576aafc240f6c87 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0775_14983643_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecc4855ad7fc6e0a7f2ac60b068cdda4ada4d11f54bc43398ace075d8a9da521 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986596_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986596_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..568b5d17f721a4960b12f0f6af316f79145d8bd9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986596_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c1f8b455b05903b8f5f522e0cfed1e10307d016e88df15c2a0e8a712eef8539 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..de2ce8051711eee3ad544a4d7b9e0dc7dd84b507 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d07b2edab130b5d9bcc4ddff695baee61d4dad3e866d1579b684f0f81194bd1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a01ce6bcbc08a7cf516f8c183f1a8dc0ee104d53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0803_14986601_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2eadfaaff704e6ce2bd739b23ead5de2bf5b0e9cf7950705665865809edf3dc2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..322d6b827a938b7098779bf116ba41a155dd2893 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91e2dfa8b1b8e0961e1317c8802d3d70aeabf2e4b57fe8271c02ffc943e03f11 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dbd07bfbbabc1fcc7f6452cf278aa87f0739e639 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14973602_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2de4f70b0628de7635a829ff35616421dd6e99d9cd91c6ef965022452685ef8d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14975811_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14975811_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..52b4bc944a68cfe98609cd19f5f04342b50b2f45 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0807_14975811_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eba75ec7a3afc9513dc6c3999be353dc1f92cf3a8f3866116fde8a6526c5e8b7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0808_15507861_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0808_15507861_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5df6795f57f8828709115d41018fb20b146de052 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0808_15507861_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b65e761150ea707384179393ba573bb06ec4f114f87b7843e969150e9b333cbf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0821_15359462_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0821_15359462_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a572f852e83698efcf825108b304e833855ba34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Park_user0821_15359462_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd6206f87176b74c64a4e02eeafecd996089c754d2cd4b1e516d26b3b0be3db2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983373_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983373_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..8060d51eaba1e8c4dbeec776dd60b5b4951887d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983373_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbb5ea14e450e506fc33f5ea697d7614400c6734a98ccc8740e8d3ebeba492df +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983392_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983392_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2025a38fb68d63f350608c355e4837d5388f7c4f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0014_14983392_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:552e3861920be7e5aef48cd4fc52c08e7cc89f1d6f2abbc22c93342a157a7023 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866966_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866966_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..19bc4d600daf50f68f9b7213d75a6eb926042c00 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866966_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e2d48341bc0512bbc4b80f7ec8408e9c6798213d537e98656d2d1f9251b8df1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866976_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866976_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c73030970a723f491db192d616dcfa9bbf99d19f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14866976_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e74cc71094cf3b8b8e4bde79717e1b709e8085331b4912f0a4b4364f27acb634 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14868790_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14868790_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..42a070a38c372ea29c1e431769156a3d671c5a34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0024_14868790_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d61c0057c4d74a9de7ffb7aaf3d747c105e7d7911ae62f70313d21c9c16a994 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0063_14867581_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0063_14867581_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..795c2e8b96fd89f2e204f67640d2f0fab7647081 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0063_14867581_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:418710fda5a328ab003f708a9006ea31f5b9500688a97b61a15cef20fb31b63c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0075_14862264_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0075_14862264_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a1d9f0cc31693e0635e7d6080a7dfa03739da9c8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0075_14862264_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6af8de73abb0b2f5af3ec2504777ee71d0ec2720fa4c0fc4194e2cd64f18d62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ed22008fc5bfe3f7f7c5ff047ca56f29167044ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8006cea38a94cb85061c7417c1712ea4012cbdc4a40e3ddd5d1b48db9167e601 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0379b40b9779f85bd27a57b3601c7be9ea9eebf9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14875972_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5de9f50f8c50164336efcd4d7b2eff28d259a8be9723c0c3f48c783790f1fa1e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895415_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895415_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6f5323b458946aa9f879b8c3b8fe22c48126498a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895415_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d39ee6c900dd3236581559a8bb71a47620ba0d32a067b6e6c38801bd3f20279 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895443_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895443_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..792626b3ea08a0d828586b45481f01f03576d038 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0081_14895443_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb23f500d0847baeb83a5a93b6b96de5adc43b7cc9f94b3637957a5bda85c2bd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867436_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867436_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..464a2cd36b44c32f3e20b7b4f4918153caf7c047 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867436_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e60e784e9e670ead9e77dbb7de4725770ad4f15fce424ff7d46f7ccfad579b0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867461_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867461_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..51dbe97e34e830228127ff6af2d141c41f9119e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0108_14867461_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0c94bf07fee167ed7ebcbfde1725c30592d1569a242cad54465a58665265549 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0133_14979289_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0133_14979289_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..68e7e7c95fe385d7cebcf0a32d37892650421aaa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0133_14979289_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:234f3b18b6971ac61d2006ce2b00463f04688fbee4c7a343142f1f1fd2b1c839 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0148_14970247_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0148_14970247_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2b6293decdd04f5a19d1a5f4f6dd2faf9683ce1f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0148_14970247_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d24d4f9d176d9dbb47a83672fa3711ee95e546feda64f834ae291c28dd5d2b1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0165_14862164_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0165_14862164_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..712cf5ccd79dad6d1f4ce68904a0fcb93f16422a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0165_14862164_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dd7b5c101d4dd2a1799abebcfc09871f51a6dd337f387e47a80ea960715ac3c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..31964b9f47f570549d48e1b2992635eca2a2538d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:207ef47a628c5af7615aff63685f73e279479eff95bde632568344b2c936a977 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..33306404f93ca734782fcf0d73f4afa5da9bc56c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984233_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e49c0f90ed6bac010c782837c28d85e08949dfa43e28fb86244f508c5130d8a7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984242_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984242_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..e7951a6798bef979d2fa9d1ab6e8b2df0f5c4a46 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0167_14984242_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c74df48ee717f3c8772d8ad1e2ba8094a459b9572b25b4d47e4763d3538432c5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0235_14868431_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0235_14868431_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f37c86981659d30661ad4bc80a5888f2b7ff311 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0235_14868431_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41e9b43d885b1ad74e1b4f7fbcd90bd449ced891fc75347938c79000c3cda357 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..5f3e8991e54cc45a2c5ec2945d77486dc769eddb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee4baee56cb6610175b15ee3a150ba259435e61eb18038de5ffac4ad8106b921 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e72d750d12671e9631485e8e86f10b743e84f757 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14885471_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45b3505d011625e95404c23673b2eac46fa774c7a10424f9bcf8e95c05358925 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14889996_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14889996_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8a6ba5226a2db9f9b7ba5a530c97724c6e8ff910 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0243_14889996_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fe983d90124c6ece8fac7753bb78e4d0313e39e2335b80495f519e7750b1a13 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0276_14868981_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0276_14868981_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..64930db6e5aa46ee45cce948f4cfc087b977559d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0276_14868981_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e719fee5991326c4106a1286ad041f289459d89267e8b87f7fca41b492a66bf8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868725_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868725_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..da3b703270629ae6707cffa1c064765c508b02d3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868725_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3034260f82b82cab767b559d435f7475ee41c0bf1c7cce87801d308199f9f670 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868944_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868944_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7f6d7e9d1f74f5622979aa7fc48d51663ad5da4b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0308_14868944_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3608d1a4b15835686a3fa1d985810c8bed5a93fd41a6ba59e53c8ba1f360cb0a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0339_14871472_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0339_14871472_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a11056e3c19c6f526b1f6c7f36ad790560e9b383 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0339_14871472_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f46f66c9e4ba527fea5013b2b452c483d5ddd8b78888d36d7890fe926ac041e4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982346_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982346_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2d3456bccb9328caa42cdef7579327c7d5d5b319 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982346_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6b0a552f0eb28aca2162ba74a35a5698009e61650406677eb2cd0a576ccae1a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982448_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982448_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a766d0af7b0d6b0f85770db73bb62074f8424ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0352_14982448_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a503ec12d58c31375780112426c093c2842355565fa724b1fa9b65dac40e1a34 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0378_14869326_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0378_14869326_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2574af513c2957999dc2eaa98602b0f3cc5c088d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0378_14869326_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c04312e594e4486f86270c258c243152f6690d7ba250bc62325e6b81534ac6ac +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0386_14863004_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0386_14863004_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..1d14aef59de6ccdbccd2e3d43a06f0183595de5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0386_14863004_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f87d732fd87a7b50a6d7f4344b29cbbdfd1919ce3b4cfc9c1e70e5eb38d37e1f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0398_15017725_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0398_15017725_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..17961ca153ec8b48f471217b7e6b09d0878352cb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0398_15017725_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ccb3ee4dc9a6ca2766f4b07fbe8afa745aefb9cc2c59d9c6680cccb6b465c5b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14866953_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14866953_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b3fb03dc3db822143cc19480ea2bbd2dda3d774 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14866953_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9539b9912c8be6bf15ea76cf3319b282a01ea015d38ab428e0047245d81cfa0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14867588_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14867588_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..cee1ef5b356e7c8d629ba41dac9bdacea47f5f9c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14867588_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:092d0f44c97c89fd209433d5b576ee192e01c49a0b50b8156d38332c63c1b176 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14887517_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14887517_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc56c96554b72b1b15d682c1892cabb682758c4e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0411_14887517_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c663690a92981bfd17814c4e9f5d44ac1e8e2557bdc739c46cd57f80b00debff +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986628_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986628_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..88660506af2eb384b82d7cd11dc01a8d8259b985 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986628_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9efebe13656ae0eeda0e169439f1063c32233aad9d2fc25d064ca11f59154e89 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986629_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986629_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e1947b2437c80112f656dc19ce6f8efbd2749ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0421_14986629_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ab2c2cb0b1cb63dd84048c202020b3365435c290a9423422718ef4dd1772b38 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0479_14869535_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0479_14869535_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..91636568f0177fbd266e8abdca6480fc28121aaf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0479_14869535_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d32fc3f9ff03a7beed6010725d6eacc61ff598f39f14e664598c72f25af6be03 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0500_14867583_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0500_14867583_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..629c74981e598817b443a25acdecdd1c50c7a869 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0500_14867583_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:801d6f26302fd7bd35d1b974091dc151ddc85a362c30fd526cebdc3bd2434aa3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0503_15016495_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0503_15016495_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ae850bc10398fe80bebb5b501020034e24cef587 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0503_15016495_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcafcf41bfbf8eaf2bdd69be6cebbf4ec5bb2745ff6740afbea1164de9207805 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f177bb47e382d7ae36054d9dfb00dc29f44e61dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bbe38970885bfbc891ef28e593aeb6543445da57986e124360867a64ef2b4a47 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d9b1e3b2b177e2c875aa7abb81de9785eb02d275 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985895_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6767b9fcb2f2fb1a6779b9a10f8f52b1f1bce0cde7aa29ee20d5ff8cd16d19a2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985905_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985905_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1035044592a9cb2cd5d05279865d922c3a4bb512 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985905_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9dc4da62f1199f17c6f90fb45332bf34258db3c9cec69fda98a22d0aef80936 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ab0cc06653b463692db530b7d7be8ccfb19988f4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6359dcba85f1459a8e7913a3d376f9d85d5b9c0c74e5541ef724193ec553298b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a11fc635f001764a58857851f274285cb3663dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0536_14985908_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cb38166ba2f4dd10ae405e3c0ced2f8b922bf63c501dccaa420f9e5bf0c3e3e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980214_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980214_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..78c92db4b4164c01de0d08e87eff1acb7933f279 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980214_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c333b6d8648a2b1b77e1a33bbb18adefd4c29368228f2e669b1f6ca459e06117 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980547_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980547_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b99b1ddc773b99c3fc8608d9c83f44d49a9cb868 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0560_14980547_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66eb9d72e4808dad327de866b8f671ade7116626d0310e3f96c81ed4e4f92067 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0567_14868353_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0567_14868353_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..01a61d7c4b5f4e597aa8fb55010d6faac6ea2e94 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0567_14868353_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4afe74e2094558a360e525c1ed00bd1c2828aca7736bb91d1be43a3edfb7542 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884600_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884600_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f89b61022ed34d8e6e0c0e6b2426299510f9d0ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884600_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e47bbcf407ed9ff14cd9ee63ff2d7584b10db31132d6324f00de7da7ff3cf3e9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884604_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884604_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b58991586a97bdcbb4f2389bd5015cd37ead4375 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0580_14884604_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e97cadf547fc35b0ceec3ec616a64b6858c36e3aabbf3df0e4ea658c7fdf8eb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0592_14866641_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0592_14866641_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..14edbf7812ac8130eee22af2e31b4da309f6c754 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0592_14866641_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72f657a7d90cacf2527adb30a66b06e3894329cb7405ab136dc7e615a73b60f0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0597_14982264_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0597_14982264_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5becb2f222e0e1c2260e0bd0af85828fb18f1213 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0597_14982264_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d656c8d66fdbaf556d7bae240d37efa81bca7c754cb83bbae0fa06be67f323a5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867959_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867959_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b74c07b42bca8305a5cda98b7160bff12dbad8e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867959_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c538f18217ce599af7e9541c3784334377eb14664ef1fb0c89d11a4605a43068 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867965_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867965_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..3826ec3bb75069066085b01495955c47d38db885 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0622_14867965_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ea9d402fa0a54968b693807f3fd2242c90c5e70be1b1cff92bf5c23962f02f5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14982958_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14982958_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3655e464226e7664377d6d362c6162d7f934c3e0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14982958_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11e7ad8716ddd1df9155529b28e694549beb641cacd1e4a051cd29fe934f93c3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14983028_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14983028_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..cf61a40bd5ceb1a18e4970d3cba8a6fce67ad312 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0628_14983028_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c06219de8a191876c231f334f174ee8444de9a15fce44d73a2396cd23911bd8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0643_15018022_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0643_15018022_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f4d38c5c7c9ccf44b8f43cee0f4379922e553b41 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0643_15018022_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1bcba993de5c0c35ee17f5c520c380e0b1d296d80c5a234132be49fa850d973 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880710_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880710_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6e840cf23397bb5fda8f31de698be7e870ffefb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880710_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56e10e2d6dbd6a195a7468af3def3bad5b309bfd73b32b573d0e45d59febf904 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..40e0df543444f6003047172ae53d54e6e7aca858 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c9b3b1dc49acb89596af9d7e9993335526864a0cdcc0762f7984213c03062ad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebf9adb652f56e941d43bcbad16cd1055d76e89f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0670_14880716_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d826d657457912432a7f15094bd5413c3958bd502d263079a423cc0cc1d306e2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ddd11a50e4608becc2e21e9c0ff47abb908fea2b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cf5f6855f78c2dd84d14b924e23dd2f95dcb38437e09f9fc3bce40ddfde2a84 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c19a1cc8323f9a9b52422245085accda11bfdd2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0680_14983897_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf056d79304b6e161a8b87faf376993bd50c02eddca32a25ff03ef4ddd6cfdba +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc8fc9e35863be18db8372955be15a4ab6d26c14 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a40c77d290ed3a5494108583950a9c40c361b18722dd993f9f848a907c466dfa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..be476594929d42f73b43eccb7235e4bfb3deca8e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31007fbf274a588d6df4b7e4c3745c021166ed8cca746c75926c5b05d7a3f2bf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..66b9b7baf432371c5bc4d4f8bdfc7d07339b484e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977450_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f85fa645e349f9409690abb59b7937670de9b6f1a0f4eab793259c765e77564 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977500_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977500_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..2f827adf91e77ba262043ce589496dd03ea7f073 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0684_14977500_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f942a99163cb77278bcddf7439c5b27bc9ec7a6f7499167290d5ccaa8cdf3521 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0755_14884555_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0755_14884555_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..56410d7b9d97723664f4437f2f61b2689e5a397e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0755_14884555_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fc4deeb33cdf6cdc5c54534759f2cac5cddd7e2d723a59560ac0921503bc3b9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0762_14876482_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0762_14876482_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..34f0bb04aa35f73c63a2eff245f4d21810e191b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0762_14876482_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f4cd8d39404f3bc10708e7236ac3ac27af234ae8a3cef1279d0d92a2f6e51f7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983231_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983231_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e34350991252b6ed151c919ab9f869c12f340447 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983231_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1874d24e289869f09df6dded9e73fde3979d3feec1e558480b4caaa8d1ca296 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983232_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983232_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..bb248f7c02e3289a1c591e0d32b61686c202d246 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0777_14983232_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:475fef3dc531c2fc3adba74c25298f9374aced2214169aeb6fe935499355e236 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0811_14979386_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0811_14979386_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..f59ff32e3709a6ece7e82c5d5a939f7600ac5127 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0811_14979386_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a089da439d140b6fea96a22ec856f0b3f08216c45e7e8b1bd87eb325c4471bb0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..25abc678e62269c5249dd5599be9c318510be6fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a2dc372ff394eb23228b9ccb6ae9b2cbbb4b5b3e4607f33fb3e1752a228a341 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9482dd03cec97e917e08d3b65d0bd4c22c824c3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0817_14984040_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:101cd039962d38f22051d1e10fc789935e91102e355c1525ba3321c089bd7e56 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..59ae1f6222cc88b96763770861485b3459eb7f9e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c29a92df0432f2a393799c312662d7736eaf1a65548f3d8d5c64ae4024c9038 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d28c69ed662426a636c7ebfa0e960602e193f90e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e94111c11bf8143d126bcaa6249b4cf3285d34140586f036fe4ce5db3e2e9a7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dfa33bfaae9d9c1f9470e6a51b64c8e994993d3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983581_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9eb9284279837ce20a0d149cad37ab2c547431432c4be02c4ff36758a25905c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983583_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983583_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..3e8d581d3546cba2a210f2412bf4bad49e26d4ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/ResidentialArea_user0828_14983583_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd2c99d7bad9b5091f8c9a5cbb61d27ec74f3aedfe46dcf824bc0165bd2ef0cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0044_15441273_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0044_15441273_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..481cb0f79f0227dc02fe45e274edc9861f26a253 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0044_15441273_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1746b9361fb737f0106c30c3deb69e9c88936a56d7a45d6835bc8a3af5207de +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..0bfe72fa52cf320df4926186d94b8165e213d65c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d81a6d11ad892ce2709622b482a279f30ab94527ad20d1860849bf0e05b8919d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..179515def59a8dd2bf49c06e7922fe68003bad76 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0045_15053600_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b901007302ace6e76a682a8e8c00d54cccf4b636da433f9e0c9cb3708122629e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311055_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311055_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..308b238699a1c97cd3a99f7d654597778f79383d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311055_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:711edecf13e3bf7850128298fa402223b1b584275edb56de48941a9255215545 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311059_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311059_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6e06671f517942d729c694695b9969a75ada8227 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0066_15311059_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f89a66ce7ca4b9ccaf7c22844f4823396129fb837ac2b3668c8a8b753ada2994 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053646_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053646_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cedf4bd439cc8f71c0221ec9a0d1fe0f90a08b30 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053646_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:941ab5c4c41401dc7f289d949211861e01a60e54ec7dc546652cd59e8d373217 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053655_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053655_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..baa55610ee6994fbb82e3fb70fea25cc4e695c8e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15053655_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b963377233700594d9e7c57898d6c4b4d1fdcb5412546fde3a1d13eaffe071c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15440865_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15440865_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..1853c63e8a8600dffcb02887caefc86af66eaa78 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0089_15440865_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:580236d4cd5789aa9c38d7a5d0636100cc83c403aed892d28dd7e9ffd0ee5513 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..942d3d6df70d2ded98b83ddc3f775e5924803aeb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d309b08679e0eebb59d74142765968f2ba95395585fa2c4ab5e44a010ac3891 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c22240e9e37dff20a7b45390f615b1bb7810173e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0099_15122216_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b687473e5ccdbf8232dc9cc9ab8c8cf0045a8d2519f2590bca47e3d4454c6ca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..9d8970e8f3b6d2ced753251cca0abbbce20d0e84 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecb60460854e4b251e2c33ae3a26cda2b7b54e09ddc832da312a89515ca8bf52 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..48305148eec59642f8f7a5d04ee728ae824125ec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988093_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca5d5fc181d7bab0169de35acc9ca16c90f0c90c1ed987792b6029eb854c79e0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988094_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988094_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6cbd625cf21851dfda2be28204758931f6613a3e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988094_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a3e0eaee2be3c662ccca4fdfedb3be9d54bb0e1f9397d3498cd8cf9703e7f62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988098_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988098_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d69983c83d8c9e08cada4c64ac2cce2b9bf41f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0132_14988098_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42e59f57a3700c4a796a49f0d88c3ddafa0398eb5530dbd7910ba9e395141b0a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879931_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879931_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c94c4ffe559547f02f58bd31533c1fcdaf74cfb3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879931_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eac27750814cfb490cbeae8a6760ae1aecb5ccc82f2963a29e52ab161aa3c323 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879938_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879938_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5cfe88bcfa3bd50bd11a25b2b64bbc19944498d4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879938_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:120f4f8078f8bf5fb5208083a52d2635f239186b1218b692e069a2e0d8c794d5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879944_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879944_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3c6ae222f9495cc36e71d199d3f8905e2232a36b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0151_14879944_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c4c55b7f25a6b8a53e96bb722616af5928114bd4513276b339780af0a7130d4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0226_14982495_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0226_14982495_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c7e69e01f12443c78e95059385e6efd86f190bc7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0226_14982495_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e7ab1fca4bc932c2dd7cb42ea51213ac5ffa3d12a7a42778d5fb51ceff04579 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0240_15519010_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0240_15519010_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd959c54759501124619a785f4388d91dad1ec0f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0240_15519010_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71372102f862c190c73597e3a97efa14658026c9254d624d2cf63125c47aac5d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0275_15519148_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0275_15519148_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5da582f432f3b3b60645c950a36c06b0c7ccf789 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0275_15519148_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40625c40cc419e5a435cf586aa6746c58bf3ab270dd81d376f841d8557e5a842 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..a9c2e3e3b5bdde7872915ffc8149ddbca9381a24 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65887cdd68fdf96936ea895273dfed5976f1ee4b9656ae46b8c4fdbe2c75be57 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f053db5198d21601a729ac4ce2bbd9e650a9fbef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519022_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa35aaaf4b2fa300a62501c1eacea622b3dbeed0b1ad0bd8a5666321e242ba68 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..baccba05d22b8e1151570aaca5971dc5c3578d8d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a1a5a0be68a6dd6fbec8f4ebd210f5da9293efc59c8051bb1161e213260a56d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..7204f8e37fa445ec64265e82539f5769a53b218d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519023_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ada56f1c68863f92acd54a1adb98124c43664cf8a9dbd58407fbf9a061f26857 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519024_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519024_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..aed36de7b7c5aef146058f939db9807dec6e669c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519024_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a238d6b307ab50528633617ae28651f63929faea9f9d137e65ca992cc8c210bd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3ba01d2336483c2f9537856df4fa3f6ceed019d7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f05273c6e031b6f5fff0c36c5c8f99cbc848f4be832c121bd24b9a87f4e22ab3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..0a6ccc4ded6fefc8f8c4d378a49a73858f5fe56a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0334_15519025_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7665c3a825dc2361a8bda8117b70c8548f80f60f4f06ec50764c487413090ff +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0384_14977863_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0384_14977863_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcbcfb19e9ed9376df4a31c54d58f0f654d0f544 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0384_14977863_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f25fbcb56f168878feabf3af1fc7c14cbe580a16873909538ebf80f61f1e109 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0423_14877986_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0423_14877986_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..30e218d33b8460efaa7c03cb81c142b6786e98a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0423_14877986_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23da211ddad1dc54cfa4de477d4f31bc6cf25735a38cbddf4176446ab7e84bfb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_14880359_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_14880359_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..9001c35709b470bc2cf990a29ace88695b72bb7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_14880359_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa3b75261284718a5a8d868d1120e6fcc0fa5d3104675e03cd0c0b5b052f7edb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_15311800_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_15311800_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..623881e5d41df42b211e265a3277b646b47251cb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0498_15311800_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32e8c0d9cd057ff8647c37c7958037620a4e481e6722afead0677673c759d3f4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518826_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518826_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7101103710244a73d38572170b7ad19a6bbe8dde --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518826_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed5e43fd9888a1c5f67e883bd91df5c245db3810ae71f96f6ad407a2cffe2ca7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518827_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518827_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..439516ebe0e0f363669125d092aaf162a957b7c0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518827_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd32f4849d58860c9c21fb040621e9a28255acf3f3bfe28190ad8fa3a65dfc5d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518830_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518830_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b88553d38f7a25e36bd32bfb4b5da3d0d7c61148 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0500_15518830_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:120bead78285bdad3b03a29c94be8c9e2bc0f45010d2f85d648c6afcf8899542 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0c3aaad80cac8708c72e489057dafb24c6343470 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bba1ed60f20eec93751e6b4e29ef3d638c3a8f764683b8123402d2927d82724 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_014.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_014.wav new file mode 100644 index 0000000000000000000000000000000000000000..30ad4863bbad2b970a8a5bb97aa286fbafc932ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987615_014.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5adfc347ee9935b9d6f0493cfec0e162dcc0d30f978faa555cb14e967b5bdfdb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987617_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987617_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..343654bdb20832883fb6b814fe4f0a69a6030582 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0516_14987617_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97c38f7ee165a6fd42b71b05348655b5d6df85ad158adf2b590f753c62b01775 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0530_14983127_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0530_14983127_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bf97f146728f035c5de4ca5e626f282774d9c01 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0530_14983127_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1303993aaa1e318eaea07f6a85c9838a0099b08a28c6573ae14c26680a0a07f8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119121_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119121_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..bbc767bc75af149b6ce78c5ac72b425df2437b33 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119121_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e647ae091f2bff1d3df81b297518fa49e37c2d836131ab508cf0bc4d54aa94e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119130_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119130_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..59936717cd090c18a626a993ae01e014d845fc77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0533_15119130_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:972bbb867249d53ccbbb02594c40229e040b41f2050b72074abaf77405925dff +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894330_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894330_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..128127a1fca7c9a211eeda5bd729251445a1f878 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894330_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31e55a335e7ae50f78188475b8d8367fd58b22164767f1b63ca72697bc2f63fc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894339_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894339_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a34a42b0eb00975dd971842e987b22d418c401d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0567_14894339_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29b1d963f30ccf3d79e21c58d6052245d29753039563aa105848eb9015c2918b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0571_15519150_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0571_15519150_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..be0e2fbc42e3c60fbf44926fb1d31ab2492e7372 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0571_15519150_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8ecb1197b4da975e730d1fb2ac9cf6f817ac904f19cc100d660eb01cc82337e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0584_15016556_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0584_15016556_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..49aa41973e9d753f31d40c9f2d5ce545a06a04fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0584_15016556_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:baab5ea1027cd34f21255f2ee5df4cc0b3c4696ddee5f29e55a882073c3a3504 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0597_14889500_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0597_14889500_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0e771a6d6ba8a1d44d61907692f3b91416b07dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0597_14889500_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88fdf5fb896be1020538d40f8826b58f04ecf3988c7dfe5024812fce9f7eb20c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311292_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311292_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..37cae1f2dfc9a709ebc3366953c1752c9a94585f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311292_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f31047e763d2393bf5af1237ef9b260dd213a83ca65932fb8ea1880169bb04a7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311294_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311294_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..69b4d2ed61258a7b4007c9e14eb421420e4ab323 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311294_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c2d1cd770236e2c8791bb2abcb8372647b230463fdecb969aa05c228e211e5c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311295_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311295_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..faca975065d96594ad65ada4a5989dacea5882ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311295_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb7d5197d9c35baae304186717a558c97bb854c0726cd3ab550dab6dac14f1eb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311298_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311298_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..294bf72225a9a05f5c83589b4f325b65e2e05bc1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0626_15311298_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73edcb9a9ee367493c30ae09106d059aa16b1738a4b075fbf2ec01fbc04c31f7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a6d1a38524a29101ab62becee0579dfb4b18075 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16ab0a9ae072bcaa4c9bc7d4540bfab06b7e24c6fadea718292c37b414d83ef2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..591b2c9c1246d276dd352d5e3769b7bf039b32e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0652_14981631_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:651610a3b1f0b36067d55626fcdcebb7e88b210cce4606c11236f0bb376e541c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0665_14881053_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0665_14881053_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a9f60218d6ef23dab2c5ae36c5b88fbd214d048e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0665_14881053_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52c6a3515ab05b8b3321753d760aa8c958319435df1a600131b830065cf06ae5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0682_14882552_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0682_14882552_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..30716a2e18d00e2caecf160b05c07ad909b7e9ce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0682_14882552_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:021351ac8f0402e4faf1a5db6a2fbfa16a2fb5acbca585e04146f5d7cfc15faf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0684_14983584_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0684_14983584_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c47953d86db68ea9e8a5749dbd04f3883a3d9d13 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0684_14983584_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea0bbaf757aae3ca1c0991f97ceb6b8d28249400c91b6078a7035383b543dba5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..0dba63f95463f4f0dca891a32c8f23e47c38c4ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c73481065fa967aaec6d813321b437a03522ad94f4158b073c31e6efadea995d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_008.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_008.wav new file mode 100644 index 0000000000000000000000000000000000000000..6ac68e5558103986a68497a7e0db11b7fa8306f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14879459_008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ba1fe8eed5c16c1d539c87e47e34e8d9214000805f61daf1d655b11eea0db94 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897360_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897360_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e14129988356a12cc68e4b5bc2bd7d8ac602f1a7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897360_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f87f1b9c54a788dbd32f17b4733e1537d3bba6cee613485a888ced30304811f9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897368_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897368_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..aad1676bf27041a9f58789f39d953d8b1deeb755 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897368_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b13f6716b9187627f4c8e13b7cccdee9819631aee72db0a702193dcf0becc6b8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897372_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897372_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..13c8cb497a0abda7848868400f70ee4712607a16 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0755_14897372_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa20606d0aeacc22bdbb3d14891f11b24fd3ccda5bc8d2875888c485e59ce99c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984103_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984103_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..df90a0bade280342e8c74dc646e29541af7307ba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984103_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfbce4aea2bb8e671592291c22338f431d0249b2e86ce2ae62bf1bb32ebff107 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984122_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984122_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..f5f20bf3f829b672ca20b254f497f2b3fbf03a1d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984122_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:831c866adf28e8357e32c2617ce97d57ffe2055ec746ab7d388f6dddeb5ccb0e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984128_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984128_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..3811c9bdd2f6e1a522744bed4cb4d3a2ed698303 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984128_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2e991b7fb4d0e3ea120c50c5004ce72290e8b4d5fe8dd5d52e286fd7d72ebd2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984129_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984129_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3fd2440cd41d7c8e5158cc49d8257458f2579c53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984129_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2f8dc26c829b7c595c0579827c341cd5c9af94a5d879fbac14ed4da4b507941 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984131_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984131_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..196c35dd2f9b67b9125ca97940abe5e7e32631f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0759_14984131_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:888dce77f1b2f4751e5872910db1f3b6e25cea74a5ee8157c89adff85ef00975 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884670_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884670_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..748ffabdbc2cfe14ad1aa7301afa151c4ec71533 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884670_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8bb7e4b331063996167f341caa612b5b54246ac8043ddb422b93180700b5900 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884690_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884690_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..6dd59d78e5e1959918fd04a70846bb77e5a43590 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884690_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8a41fdb790c0bf8d564434054a32f77706f80416a08adb5ba9691905c6aaca7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884694_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884694_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..65738ed1717471ad3be14b2eac82f4666d0fb576 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0762_14884694_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c8933f179571cb91de2e4b6b95772d7bfb16f78d739b312762995701174bd53 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b178010b56d4c3ecf4db2dabba45cdb316727461 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ff60c59a32df65caaa75ccc0b56cfd17e372d93e2bfdbdfc670a867800b69f6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..fa209b840762634d36a6084070caa563d2d60b3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_14981029_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e7b14e4a4e2d3ba1f6b528e4f1ccb788e2030ff9a7a621ed564feb965a9030a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6d5f5f0c4f5d73d5c1a39429d11d97405717d2b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:031fe741b7a662b450436b2aab4dc6b734b543edd8eb3aa72de12cd7e569862e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f68463516feca5c8d6a1db7f79ac30bb3d9c9f97 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0799_15039676_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bbbe05f72922b6cdb12df69ea37e643a79b26f6ea7ff2861a8190dc9080a0d3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884481_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884481_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb13124b835b95eabddfcf2ba855e1b85bce99f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884481_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3ddf0b59feb7d85aafbb33ddb526c33f568b80f2cd9de8040825dc1ec3848e3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884516_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884516_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..5997eb41956e096819f9eaaa40cd7efb39460d81 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0807_14884516_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b5b0a72c1827cbdb1c2259bfbba9a53aa58b0f736242d8defe3c697ad3b3927 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0817_15020543_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0817_15020543_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..de84a37618e2a34fb0f88714caac738a4e1a75c0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restaurant_user0817_15020543_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9072310bdc9dd455a8a810b078d88dd9f5cc5a58afa6d7078489f491a2e86961 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0012_14818556_013.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0012_14818556_013.wav new file mode 100644 index 0000000000000000000000000000000000000000..c10a35f2f2877b2b09f9d58532b9ee5641f8342d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0012_14818556_013.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53895158bc3600b7e3c6263477b6e1c8161cc1e35c0ffb271977753aa60e6d3a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833351_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833351_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b5edad6865ac329a95f47ed8823139bdfd7d7e72 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833351_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55f3b9f175b13fc32a9746d45b5d051dd7298265ecff237c6b428981a1a94bfd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833379_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833379_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3ac5dd083e153ad9baa694f080f47fce5b4752c9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0013_14833379_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:663c40c9fa746463f9f65691261cee8d50aebfe69397eb7b995083be4607fda4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14816436_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14816436_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..49896879cc5439b89db781d4b4273d2b54a2c8ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14816436_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2c75a0bca115385e8652ab5df33d7e31fbed1a7268b37ec85df72aa1afe97647 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14826470_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14826470_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..e4512b415618fe74c7753edcf990b41d9cd4d0d4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0031_14826470_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7c8e9bc0eaeccae395fc8a7bd42b3e8c0671b2d46eab3034be1636861ffce21 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..0aeaa983cd92ba667eca58c42c4a1b1dff33b605 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b6399e756672758153f9d042a732f71e6a84f803dcf0a5043aedd6414c99d66 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..61136b7ca46533f69479be8405380fe7fd970223 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0036_14830609_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07b3cc728ea5a73b69d614298b35a10750b7c5594c857057aadba99713bcfd9b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0049_14837132_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0049_14837132_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..8540f02b2964ec33616dde8f99369cce23d48467 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0049_14837132_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5845106913ed890bfc8e3b189e02e8660709b95d403f1dc5d8d7068e234f511d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0051_14820823_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0051_14820823_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..817052311051b7c5e97eff519cb4be150772f357 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0051_14820823_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc3d06d2103ad0b690bf1b6a38568e3624517a052f397994b66ac0cc28469b93 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0089_14816246_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0089_14816246_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..79c321da67e2dd39b648f503f380f059108bdd32 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0089_14816246_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea20aa6335bbd3efbc2b36afe2e6f7279971e5bc7b948a24a87aafa5106d2997 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0117_14833326_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0117_14833326_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4cc5034ed0c2660f517dccc698697fbc2bc0bc93 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0117_14833326_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68c89518af079adff1af1f17bea7ffe7584e9eae677e17114df49f883116dab5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844486_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844486_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..102ee7d1b84dc5fb4bcef2d36d7b7231e3cd9ffd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844486_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13f809bd0f39b94b7439b10a0cb96ad032142f32eec4bd23471c65b5196de78c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844517_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844517_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..60fc7d48a57e0579eec794fbc034c00164eb30f7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0124_14844517_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07cc1d7474c24ae2db0a45cf5be27aa1fcf9d21d330a4c712bee739856a64a1f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..62c55455332913129174fa301f1e7112e1ba5cee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3193029b70a72d5c9d97d0f04b38d9bc7cca83f5c7c986e3be5e5fc78f14fb3f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..e6109341dc26ad2927c88f194581d5e574a49ae1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0140_14830538_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5604d69bebd2cce8f61cd1d97390363eca480c7dc7db37079746befcabe256c7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836498_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836498_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..47cc3ba7c22884c415db4d408bcc885647923f9c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836498_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d10beec228a90dbd6ec91dd6c47c77371d142cb7b811fabaad7e3617cb5319f7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..313602f855f01b1a3d1dce145ae0e3aa2ad2dde5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e8d40f69788ce0ecbefbc8a26ff5c70ead90a33a391f71f1917be232f255955 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..f535d2c7227c32d7d1d287a20e86882bfe32373d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0188_14836530_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea1e71a3f1d28592030b4c9b7582b522cc7f9f5e72009f68ebd3e30eb223702d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0194_14839029_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0194_14839029_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..158db4abbc86150e041405cf382221e93556accd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0194_14839029_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d9849c2e38b2d6d93be66bc945185b9b4542b7f4171faec26b7d80c1f6572fd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14816993_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14816993_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..b5d21ef9f79dad3de67015a2e30755efdf95a460 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14816993_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e89c792fb013511fa1c9b901945df1a5f9f1a7f77178b3733b0e059c85f8a2f4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14817123_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14817123_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..267672c66c95a8d963433e87ef9593512948566e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0235_14817123_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1158fb332d3bada02b0af7a4d4879f010028a5ba777e6156361d18ea1cc3008 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0245_14831582_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0245_14831582_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a4116a25060d96b4f645c4964bace5d3344591d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0245_14831582_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbbad36fad1fc181db22247c4494d86afec70650febe541662725b3eb87900ad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14842718_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14842718_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..069cc920006777bd3a1f2737cac46604275554e0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14842718_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12f08d60a20f1001fbc6da9f65439bc59621f36b48e4ebce50d956bd28f85f33 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14847790_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14847790_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..1c2a414b1c75b3b2102badc8427ae86cea2ced55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0248_14847790_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30012a9e41afc8f959206784cb1d11b3ad7a1d8196c383f49694ff4c111af9a9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0fd56205434354b58ec33a7cbadfdca6055e4e6d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca50710a1187ab8c368056ed8f46c4d60dc6086ce167c395776e270655bc4e5d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..5fbc8fbe6e07cdaf110e19e95d920bacd988fd45 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57b2fb5c69dce217f9c52b1bab93d7bae59c6ca145d4ebdb4138f06a9871458b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_012.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_012.wav new file mode 100644 index 0000000000000000000000000000000000000000..1f90a1f27424ffe66468571e5c9e234352df4bbd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0262_14829793_012.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85618271b90d66aae9764c5efcb207ef8873c0d6f3e5239083a2dd7206203bd4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0298_14846576_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0298_14846576_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebeb3ea79a75de55912d4e85f1ffb7b2ada73acd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0298_14846576_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb53a50213a9f5ab43329bf131df074907e41c93cc7859ee2b241afa83474e5c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0301_14828230_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0301_14828230_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..3446073f5e8b9528df8adedcb455d8b0752025b5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0301_14828230_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7065be1910381ed7ee1d580d4a7327ad06138fc6892f3c6156476f8107aa319 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0314_14821320_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0314_14821320_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a267a66209a92c786694b2cbf167620395b9b4d4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0314_14821320_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2746a68bf33c180dd2bda64cdad331212b41ec81dfc8113c035c7d4097eefc7c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0343_14829654_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0343_14829654_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..5709cfa4cfa7915dd6ceae67646aa349ef27c281 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0343_14829654_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43e3dece518053706c91b8893d14dd9b54d535a0ff5d4e9a37f9a2423a44dab0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..f328b98732519283dac4c7cd556a2cd4c48bbc9d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4eea398657ede2608b2ce7957a3ad5bf0d104af69e45a8aacbf98a5e1fc3f59 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2690eff410e9a6fbecad6757c5d5f3d2f99cb63 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d60fe634800173c02e880ffc3b2dddd0c1130abb8719bfeec335407a2319f1f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..13ed3f01d599c98c178c745a0283aa8c00e350a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0365_14830208_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cd7716b23da7d0d6c05833bba08855f24dd862d3051b3921adddf5599ed7ee1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14821323_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14821323_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..730212259e2ae06edfb3b3103a2dc22a8e33f2c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14821323_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6435111ebc013df8e3e821cddaa4c2a6684129508b4b9e23d3f2f19ff80a71de +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e08ae21ba0e6c2f45b730f46a0629ebcf62c4993 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d27b610e6de219c63cbd19cb7f7bc347fe79963adf6c3e50e3a15952029c6d96 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d70841358e150572cec0b48210bb090fc8d0b9e2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0368_14852233_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5640eabd7ec9b318d531e010f6dc4000daca50082bab9319dffb75d3bda8a6d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0369_14838373_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0369_14838373_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..1acf7c67ca3bce80c142fc052e50baac609b994c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0369_14838373_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:134a8833231d478d9de01a4ca7e938e813abd679cc1001b99bb50fa2d2303f90 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..aa63c344839fd8e905861bc774c057b0a4591ad3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a80dcaab1d2308b59875972f66921ce0a2dab3274fcb64dcf0e2e1ed87eabece +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..34ff3e4a1d5794661696814b7d04f570863e07ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0386_14849434_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97449d6e9a58fdd54b794d132ff6337e4f4b576d8205a74c21feff9e76a4d595 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0408_14817928_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0408_14817928_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..7c9d9b14bc3364b8464921a2bcdee3a21a23bae1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0408_14817928_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e403418a604b537124e9cab7d911c183455596cb5895478e7453bd07620763c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0688675fa1a684218ada476119d692717ca020b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1e5d9cad3e0fa61027be1bc871b79ee202515f2a0fcf455d00ce25cad497f5b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9b8eed01069ddd182c0da28a695d40c39dc532a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ae63d22d88be792e9a36e7ef9840e0df322b6cd2f611a3a913995f49efdc9cb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d8003ac82d713730061d3b50da14c6d526f98ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843636_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a50b65ff3aabc993a2b039fbc972daab56d11ed5c53d6f77a8549b3340ece3d1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843645_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843645_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..2bc667bbe8f18a99fe80d8c868296d23f7dfa503 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0417_14843645_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d45187b95c0883b91986e36d23de5587e088200ae2c1ff249628f450fbea6d3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0422_14839921_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0422_14839921_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..0d94b7eeb0762e0e1b9c58cda60b254b42cc2147 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0422_14839921_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab78e04495081320467a477da5db526a99d0b0b8809ef570d6281c1397e63438 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0426_14818591_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0426_14818591_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d022b2a6f07f350d3ce24b6f3b45ea882f062a47 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0426_14818591_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0cc9cfbd427046b069c4e4b13933dfc185acda3fb841a6c823c95d85a74224a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..99cc8348a2e1e2bb7453bedc36b6342b6593789f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7e632d7608aa752d66071a8a0cfc2125374f4b7b310f02ba0bff979c673e792 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f3facd02f0941f598cb472d03d91230cc25d510a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14818486_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eaaf7240b90c6666322d9b94c9af23492a7f29af01879cd1cbfc48a21bde9a58 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14831184_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14831184_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2125656568ca67767863cccfcb92d348760e1a66 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0453_14831184_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58b1a0f4f94e42c4d754a6c347eb996b3f9b74385eb04484ffd906c2ccf2dd67 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0463_14844305_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0463_14844305_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a3c52ac7fb049944249df3f0c5cf3aeee4eafc4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0463_14844305_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c132fb321c0e9508a22c90fcf6ccf3967bf299fc5c6ff24ec57e515b935692db +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0464_14835121_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0464_14835121_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8e08eab720b0eebb70ec96ad3696662161b19d3c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0464_14835121_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14ed2e4b739143aab16db7670ba8835ec14a4c5bc95f9a83dca97bcaa14e5953 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0473_14817076_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0473_14817076_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc601e3407e4f5084d68027833be29b4dc5d1e47 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0473_14817076_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:986f0847d14511d4183e01aa57f96012e90602e6ceb86f14f2bde780fc1f28d6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0489_14848544_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0489_14848544_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e786661e37c48b261a0168ce48634512a4d2706a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0489_14848544_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e84e0614765c60a40b8be21426ae8ae75252d48e0dc661e5e53a693f82c3a9a1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0530_14844535_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0530_14844535_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4bbffb1921f0f24ad4249718292e70632356a544 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0530_14844535_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45d8a8775b50c262e244d0e750f4a4f35950fc0647b622beffb20227e3620ffd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0565_14835833_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0565_14835833_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..49a2540db0f9f07607c99885c05b254c9a13e777 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0565_14835833_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6007d6c780ebbcaa3c12350e0ba472b88b9ff6692d23b49cf6decb3e2f8bb8e6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0585_14835132_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0585_14835132_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..affa3c4cbc505926060ccb48f1a7b4d3a0f3171c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0585_14835132_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d88ad6e4172bbd9039c07c67166b0ef85d4ffa42462598ca8e86e66b0cf76e62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..31302eecef78277d699a72281295461707339a1b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c78a4e50b0099759b80c2d31ce377508ff125fb291dbeccd5f4562e7823df8f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..db2e4e76de2b1ba7b089590be7d9d5b03eb904b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14818937_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f555da9c404db7e37eec806d8be463261708568baa874c24d45a005bbbcaef59 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14819025_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14819025_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..118c06ef60f33294e20dc502e407c9a79f77263a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0587_14819025_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f773fda348d34c93e7bfa21937c3204592f802f65522e22adc0d84513e7e64d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0606_14823344_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0606_14823344_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e8e8850a4dc78b668af56033747bdc5edfc8a34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0606_14823344_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f8fe4693a9463660dc8d3fcf20210aee9df45c7e19ca4cb9cd6e2b57a759e29 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0614_14817960_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0614_14817960_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5e44608e79ef4479a9800f24ef03c1154053bb73 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0614_14817960_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bd577b1c0e3e7379639a959786a344b78c320dcaf30e679d135b06c690aea76 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849891_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849891_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..a2d30fd9960d1f86f05ecf5822176433cb357a6a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849891_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e71351c4324e1fe50e88a5f67dcd4d81df011cbf7be8ff039d13f9c5bb057cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849927_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849927_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..3d7d13cd22261a9ccc5f2013a1d60818ffa969df --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0618_14849927_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72a6c65245b1581edc74fff7b0075c87c67d2908586d8d2307ed241766711189 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0640_14818814_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0640_14818814_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ffa55da179a0e1a64aaab64d95507968a7d61f50 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0640_14818814_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce2d2d51fbf8e6224fa02a1271873a8f6f19bac266b5b3f78457073586f6e9a2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823293_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823293_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d4437e7c2fcfaf32f4b7e2435ff2db3a1e9d0b79 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823293_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:109bbf3e665972d950a1a0139c0d2f7c355c01b7bfe89cb6d4f30f65d8ab0707 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823378_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823378_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3fd137735ba0eebc1018d061724e03fbb02e9f4a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0659_14823378_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff3d39c9c4430fdd9e292f3215b170484971e8287233fedab60ad8c2abb1e526 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14818305_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14818305_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4670b9f2dd3e13be9e4aadb2769942bac2ef7fe5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14818305_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:059eaf9485a53e9705612dd9c11616b74ae8b5eb21beb3e08e1af40792bda691 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14823232_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14823232_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e9b443e940086b14b5e350ff239dd8254a79cb74 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0667_14823232_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3e5397468bfdc1bb33946622faadbfa12820c31d1582b8e28c8e784495b6bd5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840693_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840693_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..9935c2ff2956b462ad795369ce2322d3c055aba6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840693_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04ccdaff414792e0702da3f738aa8c5c70127e02d46829caa354f9dc072fd221 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840741_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840741_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c16a6626faba6a3c717444f4ead749ed95f810a6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0692_14840741_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc5e29e7fc4e3e6b045b3239668922b37d9cf6a06eb3f046bb5b2c7e0052c240 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0705_14826445_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0705_14826445_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdb310b1cf385d536aef84af594e39d42333fb23 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0705_14826445_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c789e5e6a200d0841d9fef8c35feab8ec2ddadf0daeba604508e32e5e1f00ee5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0730_14848193_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0730_14848193_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b59281d07d8a793c574e950deb4bfa140e0d4ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0730_14848193_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f55a79c572632ed5496a0cee03c020df32e7df853657d2f28ef16b916fb80f7f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f329c1d5892f7b04cc97fb5250074bbfe305f276 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47194ee32f298aca10f62931eaa9d762c101ab1bd99353f7fe4cfc760af174a1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..88ac82ca20ded882fb82a3857a50f5398a00a24d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14836952_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb48d698bfc1412a42a48ba400862c9fb7601deb1b3940ff1b55c31266c71c45 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14837188_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14837188_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..949f14f9d52f4ae7900b61a37cae4f63837fa39d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14837188_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad8b8171870dca9d4d362fdf3449fcdee2bfbb17184b8a101d56429b7ad1dd0c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14845589_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14845589_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..389f81ceb01811ad34b9fd86d151b13ce576cd0e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0735_14845589_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96713772f3a0a1cda671e2bf19496874c1945fb10a418f6455527eaa1681513a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0750_14816248_008.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0750_14816248_008.wav new file mode 100644 index 0000000000000000000000000000000000000000..27cacf3118ab163fc435d3fb1d5356723d9fd150 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0750_14816248_008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ecd27f6c1c345162831ce12439b576b6e5a5edf70342e80e69e09f9df4453d6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14819057_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14819057_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..f881a5b9cbd318892b1a0675e5a35fee0a8ce3d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14819057_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8a8d5eb6b3ac4c799eabc90014c994b304c4d33b2fba17d2ceb89aab885cf04 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14821326_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14821326_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6608a4469ebb669d33c41a825043e85b38148511 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0755_14821326_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b4b87c92e3c18ae0b9bc09dbeb0f105ca6d327802097d930d1873be8a8f4a58 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0781_14837074_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0781_14837074_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c81e28b0101699b4c396c05ee2815aa3e8745651 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0781_14837074_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bd50ee9d91ee2ea475bba2af14318f053e00687708ed1c3cb5faed02da6f55e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825722_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825722_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9485576b4bcd703d8556537f61bd3ce123365c49 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825722_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:374056fddc784b414104f1fd99c8a57dcc0326601f1670476a5cba84d44a4231 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825848_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825848_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..2cea9a187c9d9ce22f96925dcaa95ad83784c03d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0789_14825848_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:513463f52ca4b66ef1e523dafd37822bdbc05079947c91120088ba4504ba3633 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816886_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816886_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..d73e6a24d42cc23a65ad8a594b402b5c8d78dc51 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816886_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d60945eb460d02f9914283a3488c611e9f03c44daeb9707022d8587630e8bad +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816953_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816953_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..c69d4afe8aa924ece5010de3ace483bae659c061 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0791_14816953_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cea7f136b1525e0835c616ae223414bc0f0b60bd5ac5a2505dd6dc5bca12a9e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..93f5a7c612535f9262d6f60b177368422582b468 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:365d994f32428297a3aaf291fb8c09c9e11f2ada55fde1c442d10b9c721d58e0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..57b39e4f4d4a5ddc74aac17ca8830ded441e4889 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Restroom_user0811_14846831_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba03af818baa3507a341493fb1bafc0f6a5c15048e13d703013f01432306415f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0069_14824730_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0069_14824730_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e67b4b7410822ae0dd4ff7573f1bba0b00e615c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0069_14824730_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f4eb3dc54f3fe1e29fa7d76a7f24f13c6d2872118f96d350746059150f2e558 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827671_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827671_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ced77cb29ea927e2f2b4ac9e370f28ef575f97d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827671_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9211a798efb8e688bca6a75d11cd897d56426ac84052ec86f159b36792ce56fc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827711_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827711_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d19e31f8c2b5f8810169d2bcc2c9dacaf699a51 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0106_14827711_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09aa409ffc89aa6d44d334cc88bddff120666a6cd3e9b41f2cdc851749e97117 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0108_14840677_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0108_14840677_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0cb119b7fcbb378143c49734b4b633b41d7eeb8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0108_14840677_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd9fecc4cdfcdb0397f8d0b17a8788d06c5ef598da3a487239157584c36b2819 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..3132df1cc34e88bf048df5359685afeddce34e77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9efd05d5d3643561600f4abeeecae226f1a66076560383de6338210fa5a230e3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1377403b85497274df577ec6867c01aec271b3d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0149_14827904_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66c633e8934167e3c4823b5a95f3de606f68a28ccdff6b3cfabe95d077177385 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0181_14846735_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0181_14846735_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..36b450daf3ce1f2e7cdf08579a1110e3acc9f37b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0181_14846735_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:250bac5a94393be9afa911101b6c28b5b3cbc4d5fa012c64cae5fa5508be8827 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830290_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830290_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..907a210f4acfbb63b5bb416f1edda06782160117 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830290_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:926455145f03fc88643f4b04cbe77c2c79f3cf6d888354dd4666c50374e3018f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830440_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830440_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..434dbfd61c233e175c8c41f70ccee53ebd9d47b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830440_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dc89eebcea741d5e1ad791fcc015dbe37131fde746414dfa4f2043d2459fddb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2ee0164d2a8fead9de12fe49de314f1651a9dcb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6743a8e618fcb345d8f7beb404c2772f048768802f7e7ed6dc8180649d10624c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..206d467265cfecff06f72d6e853ae728cbbb901a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0183_14830598_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50a005b7f77cd06db4e3f4f0eb86e3317a20f857362f7a586618c1237203f18e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828083_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828083_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..87cb4d993cdfdf1e5ac05eb2f0baed31bb3f808d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828083_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:442d13ac571aedead556725555923933d22d9b0ad5e6890d0ed82c22b1ce45cd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828151_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828151_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..733278f78f8ededff47f51dc67b4fe20ead98f5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0212_14828151_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fc3bd5e16de220384c8ffc9c8fe645b616a97f2dfcd846814922928e6183cca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14827881_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14827881_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d8bdbf8f7430a5c57cd31bf36fedf4badc880492 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14827881_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56b4c8b235a97eedc008c3b2509f115209c17042ef32dae4147a981c061b3adc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14834803_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14834803_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c6bd4bf56d71cb51f65305ffdd04a5f9b9c7509 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14834803_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92a43dac8bd93b3b495163816507bcf246f5b7525cf7ca32378a5e32fe855226 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14835163_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14835163_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0dc4fe655ff4ff741b0db5282ba988e4b1cf0596 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0235_14835163_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23199d50fac524f893a733d459f5d7b8096adad2290fbb3cb84918e3b74efb60 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0250_14827294_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0250_14827294_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4ed27863de9266ccdfa07d84287fd6009262862b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0250_14827294_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5529ccfdf15d21270330d97b9e39dc6b733bf0400b87fd9ff02f35db222c7d2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..91b9e3e821c1301c2b0132691c224255e558e32d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96cea3f6559e11ca304d2d8acac11513dd847b50e4091dcb76cee243992e6fa8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..801288ca6015e33f183040220c6514e850149424 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0254_14823497_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:693cdfd18f6bf57ea1c6b03b013a2cc28bafdf85e6b294e5cb8eedeb9c0a43cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822378_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822378_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..de8afb0e3ce65225ca3831425afdbf0ecbe6f799 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822378_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2517109d734d244f0faffcf8ebaabfcc6d35dec37f56af4bd61c92645dc7e57 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822457_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822457_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c941887a5488d46e6be0a42e136ecd2ca444f0e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822457_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb198aa1cbc114fc2ec421477c2fe82a1fb1cedc897dde9e43a339a75f502e24 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822536_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822536_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0878a13a7a96349837c32558f73c6d77559dba4e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0314_14822536_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c08de7f9bff6192a5fbf98634243a9cd31cd4c5cfa54562e26f7101c2f3c2f97 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824951_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824951_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e491368a270ea6ba6c72af9e8cff67c944045446 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824951_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b681dcde72e65a52c4330c7fd43ed635d4a88cff62ced5a4d4335e253f755b8b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824996_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824996_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..69b304f4510c202542651c07359b8f6547937846 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0346_14824996_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40488ffedf880a30238ea9e85088337a2f7175c70cb0a3dce40d492928f1287c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0369_14831719_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0369_14831719_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..75e941865b84fd6138ff4b57b84dd4167ab8f5ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0369_14831719_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a92c62cddef42224d6138c4ab28f32dd80aae51f60009a77f9bdb2e173cb3ca2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827417_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827417_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0647cf92cd74cbd5376e13168fce099175c2256b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827417_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0da2fd323442583b86ef763999260a77e349837e78828862c94d0518d4fa357 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827511_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827511_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..16aa6d40b3651054faf3320f6a99733a69b4c739 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14827511_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:636b8dd7e1b4e922bd83531548ceb4f16f637fd086cdd819a3b81deb4953d859 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828643_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828643_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..04ce3d57db9b97834c89fb913ece7282ea9fcd1d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828643_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2a354ea6ba6c6a4b43377bf727539e02e6771a187ea86d72e9b5ee590f157e7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828691_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828691_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..3424646ba2dc6de95249b4831aacc70acdc23a3b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828691_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:071cb086ff7f30ed62b125f365481dd10efdf63fdd8867d60423645dd7194348 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828754_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828754_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b6773415e8ee6c70fbfb5edc279855e7bab10b8a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0383_14828754_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd44b88b36c93a999a7cdefe652676992078ad0870bd385420c235526fd73d86 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826521_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826521_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..3f39f4f0d8a0eb934753346e167020f87a38d44e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826521_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebf9e4966f9441be17e9e3ba637652bcd28d887a9b11e16cca217478a133c90a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826554_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826554_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..1021b32af52fd842e859f009229aca48a801be0c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826554_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28d4e44ef6225e426d5d13ad0834094bade21d71291c70a485077a005cdfac0f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826629_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826629_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..d1560b88ea6a4c31f707049d2e723a1b3222eefd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0408_14826629_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f6d786cb9e1268ccc3b629ac50b63e7cc1ae936c24ef60e1b9c57316beb04f8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0417_14844681_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0417_14844681_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d487c0a61ec3c97396ccc4677523c5815343118d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0417_14844681_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53c8d10642d9b7c7f22ad3c7fe4ac800e8bf8b311f8c369aa2ecef85c831d0dc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0426_14844411_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0426_14844411_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..29240ecaa506b002e35a23082d21441fd9cd459c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0426_14844411_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fa9eb657efd6d581f2ab2ff0d43d92248cda669c3385a77b112578e0837bf9d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0448_14852420_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0448_14852420_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..5fe4388c28bf3977cb139960aaf563b1231524bd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0448_14852420_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20e910ee64b4a2ca547245708a5430289fdb515f20e663bfb0c1089dc8e5d849 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0462_14827358_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0462_14827358_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..76b7667870bc1e3faf0703cfb2124516627105ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0462_14827358_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62a3797ee72d5f7fbbc2850534a569ed682340697bed6119d0c7c430056fd1e9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834738_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834738_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d23886f0067363cb17c0ba4cdb65d20cf32b4d5a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834738_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a58e01b93949408219c70d9dcd5b48cd8afbdd9d08b003ce72dde9a63f4910a2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..7216a6bee2d4601b725a90b2eb29b4b05b1603e7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d143d5c1dbd5ad97b2fed6dae833affc5e0e0bf0153a9b0c8f23a4366372076 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..f3a664d3979792dd0e4b96464fa6c7346769ce07 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0465_14834748_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e093b22ba11408aa2223e7c7f4137beffdba1a35bb6d85b2c0f386ea02a70c5f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0493_14821515_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0493_14821515_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c7057c167a493a7d9dce2baf00e67da34dba0969 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0493_14821515_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8c1ff0262a381a066af9dcb3b916ea3e2144fb2c0fefec6dd2de90770fa5d11 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0524_14835809_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0524_14835809_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..d0b483ae699f1459f4a60620cd58414dc66ff7b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0524_14835809_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfa126e1e3662e6cc8f03c7d4092489d514f76d233fda36ac73eb2254f8a4018 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0530_14835161_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0530_14835161_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..6300bc32788f25dfd91ebc2dd1964d85481bcd5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0530_14835161_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e9a3a1d931c9ee194b8ce951e86a470fc367631f0b9fcd94a8ebe3f6c6b9e2d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..398edec30081b44a88f18f83822b6e0aef31a207 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f05e7549a83fb34335fbc48199047870f5e797093c0bfba90e508fa2e632f917 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..36370ac3e768e6a0c93cd38644c0ae15e5521121 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0575_14833204_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1164059a4727be9a6e7bcf2d0789d800cd57d3c20a342d71cdd59d4996e18f19 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0597_14823227_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0597_14823227_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f79251d12340fccfdfa599493b0c41ef62104731 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0597_14823227_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0696c3e371361019e3cdad31aa0ccd751caf7819f46743ec7de0de19a5a981e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14821785_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14821785_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..7aaba64cb6b8c07195a9a62a863ac735b9dc852d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14821785_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28f9dcb82f161286359dbb51668bc6fee2e75b93057ade9fff46502190e47644 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0247894894d24517a72bc1067f55e92f5f1aa8f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e83b38243c92f3fa36f298928635be905aa402c2c81327904f066b7987fa60bb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..8254991446737308cdf2f42bf3e775e0bb60cc15 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832917_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f68a9aed63c352a306a57b7bb08d1b1c3ed0fa60fa0a1f5210560d5df8971101 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832996_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832996_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ed69dbb5580ce7d2bfa26623e731e56f394d365f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0712_14832996_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6551e3b2f99968102829790626db6e9e5ff61cfe905640c773454f4f463f0c48 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0718_14831352_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0718_14831352_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..85a3a67db77f604a25d123283057cb1208858961 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0718_14831352_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5946f6d20f19a7c8f7a3661bfed213e366670932dfd69a73c7150e0ed105d0a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0761_14825328_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0761_14825328_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..78dbfc63134062472d9c3ac521cd659adf0b784b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0761_14825328_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee0f4ce3500b16937b1075b2e2889f10d4cfdb58a7f8d0884f625fa2ff841be5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833417_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833417_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..ac3f23b47d0549422680db05e02bd4c8282fba9c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833417_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf303bcdeaea7df1ac61286a26c81497a760916f4f8da2814c42e0b368a1cf9c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833514_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833514_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..5e2e7b4d2cbd96e312f372667bfd21fe56997569 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0769_14833514_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f31bea46aa7a8f52d0e2148c8604f31ed6d1877bfb5d857e3206623280904489 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14825445_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14825445_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..865b4185ec026f31ba70efdb29057d23357f09ac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14825445_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b52d5ed31599b1fc1253e429b53804d828ea63f020a45f9df73c09080cbed578 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841151_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841151_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..1d99af1aebb02f028607f2951ef6a1dce259bf2b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841151_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ca8bffaadc3558a0aa3ea84b71eb0ec661da68c60fdb31a1e017f9eefde3f06 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841175_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841175_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0cdba6cf50bd646e02b8309c0ad08bcb34f7255 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841175_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96694151cafebf2a3922ab163c41f869adbc43fcaba7a1cb4143ad174c33455f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841195_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841195_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..80a6ad46f82dfc9b9eea81f45cdf123961b0daf9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0789_14841195_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:109646185b0c3cd56f69a7cc393e2d633e056104ff7e163b223ceedefd151d42 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840616_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840616_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee9d235e30737e5f467d46dcbb68914144857551 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840616_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b399e7fd1751cf13a1edbba11d3282a501d39f7614e739d977b8f173b246ca2a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840622_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840622_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..5112e901074a42045d6a9466f89bbe69c8a95e20 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Street_user0817_14840622_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3dcc22f9b84b30ff137c23d7a05409fab471c65171dc85afb56e72da02bbf00 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a29a159b192defd2e7b88b58cef24321317fbd82 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e8c1ec3898a20f353fa2b167b8e88aa0a824ac4c420972b9ddaa5e0a9649f97 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..7cd848d053913c3ecab41e5d3fbfb96bafcc2027 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15208798_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2cc68eedbc1825c711552aeea527d1bc29bfc64f5e62e33ffb75f7eb99ebc56 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1955e1e59e5cd3bc3dc886a6d76cde494704392c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67f57a986f648bc630942415c0d23d162c70cf9db6fdf9f542e18ee6af5de034 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2fbd97abdfd839da037611044204fef74af425df --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0000_15245944_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fef03b2a772cbc67817b594051b1115e286a4fb8b43831b8aede18e66c7a993 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311279_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311279_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b9f721c7157a1de2e3d40a581daa6a25a41239b6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311279_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62c7b4c04856441b53c0fa0ac767911514634142ffaf78bf1af7666163733f48 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311283_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311283_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a052dd868969a728ae11ebe0115267060f4a693d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0017_15311283_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80d8cc81eed1f58602f764ccd3408ef42fd27f615a534b5efa7d5ba7f3e5ccfa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15441278_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15441278_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b4e2502f777c0e54c942623db2f362f7be2516c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15441278_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fec18dc9f3421841fe8869303281e8e2b14fe1541f66d41cf759258a5babbb61 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15519156_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15519156_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..c9786c92441ec57e4591cdc970d99dfb90ba3fa8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0044_15519156_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0a3b8997d898d39e97d5de823a8f46aaa35b7bbb0b2d01be539b66aa1cd3a4c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..4770762e03fd8c400cc66e49523890ecb9db67f6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02cb445850a8997cb25d59a0b0b12095f133271f4b4f3beb6df8bc9e8248b872 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f05d6773a8013809a2467ce9785feded11cc996 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15119578_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8c83d4b6a48f648be6e6e6b5bffe901a894dc8e9cd07fed230e1abd76b75e5a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15311828_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15311828_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..aea1338f9f84061987beb92d582220e919877392 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15311828_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53f9ca0a2f00522ab81b7d3de72bf183df30c1889905abee3ea53a40351c7f46 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15358475_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15358475_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..96157bba749043594475926daa7d219eeefddc08 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0045_15358475_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6689233b8594fcdfb2a500e8e9dda08ecb3302278661c0a7dff441f67bc4f1a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15354072_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15354072_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcb6a5201b29c7d27fdd32c03e1c0dd692d7797b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15354072_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:925635c821a108f5a2cc4c73d2c0bdcb30ad7e3efa6091500965c332d3f55d6c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15518744_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15518744_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..2581b8753f5bad9686f061995d665415ad8deb77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0091_15518744_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ea9b06cbe6650a04af922e1c11f0fda69e156513a92cb3161a18870f23f905a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0118_14886512_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0118_14886512_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..eea06fca665a7b8513a5989d716bf8547766a29e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0118_14886512_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:281ee826eca39fc0f341949b5f7c6545fb84edb9fc7cf34df6e0672b480c59e2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508860_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508860_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..5da9f2ade7f11a3ba840275cd77a29a9881972cf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508860_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e4cd07e9ff5471901c1845874ddb90944c4a7f0db9e63464442dcfd7b2e3635 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508875_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508875_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..3bec2a8c5b618b0433780ae890af50437705a615 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15508875_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6dc567afda778a8333575d6460eda18113698bc3b4a199f286991de44f7570b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15509322_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15509322_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf8b34d0e4107b5b49d951434f6b51c048141664 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0132_15509322_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd75d0e68d8ce3be727f2e88d312e08a25d4023306cfaa3b9f0d78867c676ac5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..0d5f40f46e58d89315f609c143746dee785ea8d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08e85b59fb6718b787e5a813c76c99cc5e96f77617ec1b20f9a54250a99b9a2b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_009.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_009.wav new file mode 100644 index 0000000000000000000000000000000000000000..a57f426b42d09be4f838f46f2e75ae05b31ea3bd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_009.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3cf4f47ac3574209511180cd798668e3db488a2b7ce01616828e90f514ae508f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_016.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_016.wav new file mode 100644 index 0000000000000000000000000000000000000000..51b0c48dd60fe84d22cdff61847ea449b10c852e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0138_14984630_016.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d0d27f1a8365679007679c2bb456c980bdee9b89074fc28863839f287cada45 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14898228_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14898228_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..502abd7d65b6358d1f96537794b889daaed423fe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14898228_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c80a16b3f822f46f0afddc23f80d44bb454b947646d5b6679021c403c2c4d531 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14922311_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14922311_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..1c57da1eb0e43a15fa7cfff729e0832466b56396 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0148_14922311_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43551748cb17dec1d446c9dd4c38479e40c5c6302d5bcb72b41cac99fdf1be90 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..89e396f3bdf25a72b227ddfa00e6c244ae308908 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5f008c793c0d0814103dee4b6385fc5e3f7d23c7b39e29bb51a45839d54f6f0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c3f3e7a49d57d89d7e8884d9a61246840fd407f1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0226_14980893_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4e5128c6391ffad655c477de9bfc851345721aac925ca016f63194c20b81b76 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b625621b24148102c6b62069176e8fea7d5069a6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98fb827cc39f2f60790f851b7e8c3aa93b24637b4965bc93cbddecf4b3bc1ed0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..b9017c84f49360a7da1cdc8a83968a868a20d77d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0267_15053357_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:711e4610cbfb81f918823a634b72b0e040d18685baac0993d5aed38070404572 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880115_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880115_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..748eda21e2411410b3f01f6574180347ae1880cb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880115_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36f7e5996751d69497eccb4775d5b8a47ce493ec9f6677ebb38221f309bc0772 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880125_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880125_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2585b9be87b7cd1d81c3a147316f0c51bef2b6e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880125_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a61c3769c10e767faf1270001343737ed623b5cd9753650c221c4fc69ab1fcb2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880134_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880134_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c97024a02d129571c33ceaf690c31c6f04fb5267 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880134_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87aaaf098c32aa75a35309f834eeb8214963f8363d2438cd461b9ea144104ab7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880137_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880137_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..60fe0ffe93035c9c611d5eeb082bede1b4427004 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0277_14880137_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92c2eaee557a34cda7166457c7d6f95292fb5e72607eb86ff32252889e804edc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519808_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519808_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b3b4bd2d415593bf996bdd88f79482ca9f115087 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519808_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de17c5cb7e2a09a7320b41340bb71a623109938416408280a8654dc726ad284e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..3654564be309261066dd1cbb8a5f591c9f80af57 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc06183f0d7f2c002c52d078a15bb61d712faf4a2fc6fd9c04d9fabf756e628d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2bcf6cb3e2df5ec7151766613d5e43e648fa7cd4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0290_15519810_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:383e6a78cc77ff896b30f4c46b0fcaab2fc299998af6ae8895b7ae881b82e6ee +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..dff576e3b8a48f04174939e2579ff28edbd2836b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60df12bde48f432bfb2d2a18dd1e5fabef088837f8e010a74dff10db6be0c8a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..25a2e1372d9182f2e18b14c84979e5e8d32be1ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0291_14885689_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3f80de331727744d46eecc4e246c8eefaaa33df9555c15f52421f354d799ef3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0325_15030729_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0325_15030729_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f54f2edfa70cdce7dd55ff313e86ae99de01b407 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0325_15030729_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b3287600f32595b75c2fb1a6043da3b4c92f419ec5ca7ee94a73859273be4f6b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119165_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119165_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f7772bae3d36362ce2c5d064839d7a37d2e467f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119165_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d8126e99a4244a6ad6d38f00c189b7909f35836670d9a789e78e826e10d687d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119501_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119501_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..e814125f1e7e1fadaf78eacebbaa82d841c495d9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15119501_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fcbe862ba8ae38863864d6de15d6e6bc014089e032f4a1a6318d5566bb53f36 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15441268_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15441268_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..57911b2b127578dce7a9ef0fb28a685f9a307aa2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0357_15441268_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7864b22f773e74034d7a8b99c7010db5d4743b344d13780288bc93bc1ff0df67 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985979_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985979_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7f49a68343cfd08ba67246628c6efbded916106 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985979_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d4a258028634c4cda642da2e6a967ca414f20583b0eba9f8431b7fa41af7cb0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..997ce67626fdc605f2d9975b7398cd14cdf83f94 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e11dfa7eeae3fec57dac9035dbd8858f3199bdc0b763fbc656ae48ce7060ca5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..a25cf83940a3162decaffd60f6e112e507d8f8ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14985982_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b34f0e95e06603ee06b0bd542c555d86b1bf8cdf131bcdb78672e3340850873e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986009_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986009_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..07227ab3ac89c7ff2bf110426c8eb19c0af39ec8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986009_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:821d24cd41770193e6f006b0f1bf3999494d9b1b41e47aec4239b6e866bd00a3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986010_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986010_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..0804e6d794b425663213b4ad26a42c38358b8295 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0378_14986010_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a35779f4cefcdac88063cc1ef3fb84ca5e08ca81e3195db7981860e49ee7073 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0381_15208819_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0381_15208819_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..559e976564b9a9812aabcea112284a5cf3ff0049 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0381_15208819_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:194d5a8dcbdc7adcf9e690341d607faefa62df314c80fc8d5dbed33e220e0d9e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519832_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519832_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a5a9690c9ff3c1ad42d6aa0000377f388479409 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519832_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8ec3dd1fe52590e1137c8e2754f0acc37b973c318d01f3ee02ddf64eb9b0f0b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519835_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519835_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..3c69dd6b9a986dd23790e21698fd05ca6364c775 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0389_15519835_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f048b5e1003d529b689264a2b3361e271f9a28a074db8929d3c0a573fd0ea4d9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0395_15119410_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0395_15119410_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..447be32d9168e338973fa5335a3823e04e559af2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0395_15119410_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c8a440dfd8942342a6c4bbb1d6fb8c44f605d8a093a4e89c0f397bd902b246a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15441334_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15441334_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..883aeac55f52d5210aeab4bfb3dfe30a61a477bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15441334_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13ca7994ead5e5ee60594cdc7ca0681f4f82187b7e20b6e5a25527ead2a56893 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15486004_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15486004_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d4aafea0b2e19b1900e418e7865f193ab61c27f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15486004_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c6495e9b7778d728a9e279a1325d83b81995e5006d7b5af4efaf226d5a3582a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508121_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508121_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f750a0463c2edbc73cf8417d276cc258203be0b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508121_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82bc65ecba0aee604b55c5836d0c1c9a29a6e56cc4648f3d01aaa3c9e7447a5a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..1140bf078285cc02f346a759cb71423a993b4e73 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:916ce78af547e81ed9916f01889f2c0fc7402d58e9a585ebd1d2eaf2076e9dcf +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf684ca8948dbf8a8f88e4332bb3c6946851cb86 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0403_15508250_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aadc7c9003ee030892d4e3fff7d612749dd6504663581361fbe0fd8b385d0674 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..0bc69fe1c002fdf34e875ab634f0264eae2e535b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33965a7c91cc2508d81abb4716a5606b697e0f96e7b4748b718967211a58629b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0b408ee8418b295401e6aee3feeac6d08acc88d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb0e1c1640e5c82b909a416babf9c98d0fbdc0be74e4c69c966be45c3a5850fa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..5000eb43c9b30abe8be8b4d65ed2d8c1ebbdd556 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0416_15519120_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:486814f957e93b4a304df250d8a37fcf08868e5cc71e3a7a73f19a658f7ddcdb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..3b1fff3c67c1074a2c90c0b18ef47330d65a8583 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cfe855f80e8df01c9585e8e4f755167e91cf284ad2d2912f07572a05676ac8a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_007.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_007.wav new file mode 100644 index 0000000000000000000000000000000000000000..c792bfe4721eb524e96c15fe3eb68de7109bcf71 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0438_15119086_007.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24a3ccec8945b4fb5d556647fc0df710d8bfa665a68fdb773c118ea3d21ef79b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_14997176_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_14997176_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..766fb0f93cc7706c32cc323611d3833d59251190 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_14997176_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8ccb2a30475dc654a981912dd5e49cf043853922844f02aac4f27697ed0c46c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022408_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022408_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..78ca2225362d21f9a8f7bde1e4747c521d80fc3f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022408_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b0073c6b3f07cca7256b551a530b04d59ecc0c546d5ef99b952d954f29da10d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022794_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022794_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..d6df5586f7a795cba8272e421fc0762a91bc947a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0482_15022794_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4db1a4544ded2f3a8a7a9198645ae228bf0247460a1db6b5004968a580ea39cc +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0487_15016470_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0487_15016470_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..006c6e95b2496faeec140ecb80e6e8d510ebe119 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0487_15016470_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5f16c5bd5654766c2e83ae9ede8d4f713a0fb9f76392789c78561946e270fbd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980285_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980285_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..e5c485e029cf5a5edbb701203058ecf30aa26207 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980285_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db88885f3b2998586c0107b1682b6cfc1af0a883755333e739f50f61ebd8f5d9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980286_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980286_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f33820acfe49381b70f3b3a89caf306342e4de0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980286_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a605c45e34da46810068accf0fb3415729eeed871b9336802b3f410bad691e5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..f86cc83913d5a59dba068380fe84302c9670e967 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e045aa2144cc2412b2dc437548bb4551543443693153703cb6b9a65b7a3fa8c4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..6f7e09b944e26cc1d7c8a720012650605214e9fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980290_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b14c2666e9e4bec4e0002ef7f172b7846a547e53124ddd9978c6a81e0ee06c8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980334_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980334_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..82f1ea3e5f06976c8ae81bea915723c113ecacc9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0528_14980334_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:735a913d7fbd41fd6262c46019218a59ddadf2926f6c332b138bd20bd76473c9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0542_15112838_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0542_15112838_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..5e71cb8e8b8bc774ea3237b60a9bc5aeb7aefbf6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0542_15112838_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da70a4bc444debd1d13e49ae9e26b30d5e865e0419707961037dcc7d60e01c7a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0586_14888949_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0586_14888949_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..0aeac82bc281d1cfb1f0f6ad853b1bff8c2c9f48 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0586_14888949_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc2e9a6a51067ed5d351e2b17397a8a8c5d87d800dd27962787a343b7038530d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0685_15119584_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0685_15119584_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4bdb1bc24f109e07456e4954aa8c051d02d868a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0685_15119584_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b96755909c068681a239dbdd7f6488aa4369cba0021e5feaea4cdcf9ccaa426 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f59250108e1b9fb923f3d6753f2f3928bb8d837 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94e7b9e7a9284bc2e7cf2776d42ddfb7bd793fe8536072d10f877a5b231b393e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..771d8c19ae8c13378a3fedb4ec78cfbf494831d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15507908_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39a882ffad5b873e991cc187277570a7bc2a44b8227aa99c09fc4c496423ca3c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15508122_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15508122_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..ac425357ed393f58798b3741d4eecfb587db90f6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0698_15508122_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba4439eef1edbf853f5f89848c2b3308722f9a9101af31abfb5188806d6e0a43 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0709_15508733_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0709_15508733_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..46a362e4af324eee4c94a9ff7ac7bc9b8997a560 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0709_15508733_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35e65c19a0183f3cdaa935a683214d980e3b4b9b422cbadf9a0f24bd872aad62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_14979621_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_14979621_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..d67a992252d3e1ccafc81be88a03b8c081dd2b99 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_14979621_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7638a766d6bca647bf8ac6c39fcbd9b94c1f29c245f1fd97b61f7c05c70b9e00 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcd82c75f7c77e5c928b1f299dd9477c25376770 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1f98641c185a918f5a7580b2eed0a33f88b3b8a30e5a6dd765f95b7c3d79d7c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..5a15e4a59f96a53888eaa348703fe284122b0b6c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15011492_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb55d734c40f2b4739482b76a26d910c192febd0627d6b03f70b355b5570c8dd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..05c1da7aaa85f4e3ee36e811b649488d0bbbada0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d0593e2e5a3565d1d86be867a20e3de6975775bb1802c75e88e9e606ff8f7d1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..19464f8130ea7dce1aa561381e86a9ab677f1e0a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0799_15013024_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50984184a1a08ea0a96f73962ff74f725012dcc9a6c91afaf29621f127af141c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0817_15054416_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0817_15054416_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..be76e9ae2ffc908ee83659261fcb9ef5eae5ce82 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0817_15054416_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34b0098efa4f509888dc13437d8d800e080ca460d8fa91f590706fd62f2de44b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0829_14893713_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0829_14893713_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..505e44b4a78f2f59afdc2dd4d94730a7e9431437 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/SubwayStation_user0829_14893713_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8748be8d8982a0c76e075dc5ba161204eb25bf7994e31906e6705bbac8840b2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870435_006.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870435_006.wav new file mode 100644 index 0000000000000000000000000000000000000000..2b1a62753a8360b7304a89a00337f86b30cb4be7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870435_006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1d90054953381d137408cb893b8ecee25b02b30f3cda3fe0191256176e17cb1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870466_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870466_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..8326d3c7456bf658479e67df366c76791e56d3f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14870466_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a7676a5274e9e201aae04c779dca4eff43164db062c0643e4dfa76a97b54bd1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14872486_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14872486_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..e38ef61d9ed357abb7be5d36d2022fed7bf94be7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0065_14872486_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9903bfd23757a53dea497349faeecbd1b4cc0b116c11e3ef27f88ea426ac3741 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877130_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877130_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c86ab8c2d3bc8d31a8345d4293f3e4f71d625e02 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877130_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41e87c4cab6e94ec1ab31d744e09a6924134432c3b83a4ffc0a6b98c1e4c1316 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877179_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877179_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..c4a6563fde9456d278cc59d2fd457bab0dab1e12 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0092_14877179_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d19c7bf6120e1f30f4e8865d435b4efd6ced1537c01182f05d680c77e48393b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890379_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890379_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..06f68f9164901e6186b8eea9abec16e8d32a430c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890379_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab2193c22cbe886d8cb2cee2bd5405e004a40cf2e0e575af6cf4f9d4823465ea +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890420_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890420_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..131b6ebd67f83a766e64462fa1dcf12df04d7345 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890420_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7dc44d12026520e1193dd28bc68378907c6149a9599473f09faf7a95ba5f3b2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890449_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890449_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..4dc7e055b991b617471ffc1efe1565fca4033fc7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890449_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da53d5e9879baf68bad58618d096bbabcb7df63ec01a42c181f7988a167c03b8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..78bfc0c3bb0f657a00cdb0e762ade11ecdcef3b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11109d7bf867e6e8740a2573d20ab54420d302b6224be584b6f4df05f807f176 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..4c244ba5464c9622b9fa8b24b483b7726ea0be01 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0109_14890467_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:12f59bc02a60468aae447ad106b0631b3a3a4ac5994addb40d755a631e193b82 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0116_14895124_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0116_14895124_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4ae6d7cf7190d60e020deeae672717226cca88a5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0116_14895124_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c58cf4b41ba7b73ed3c99f94cfb31066462aaf332619bb8386bd94933e47de79 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..898d36c5a782093f160d21b547eef036a7b2b9d9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:233cd97a8ea0bb74c4becd56e41b75c8f4e89b23feffd3029c99193fd4131d62 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..a651deeb529d49bd5aed08907c24c0361a1d8c7b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886486_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e0e45b4f9991738063636ce980da75658e0c53d27f06e33c5a2544366ec8967 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886491_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886491_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..8456c3a559496b992e2db1f6cd7bbba98fbd34d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886491_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82eb6998cbdad68c08a92f07594fda8d5707c57c10392408573f296a4eb31d36 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..be2c9fa0fee032f872ee17b5acf3816ceaadea11 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73da3cbaf2b1f0165f5a3f2dffa4d00c0b696c42c18418b7d540e60eff42e6b6 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..05952055eb7288c00c1e2da98d892732e5260cf4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0118_14886499_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25df6cd09f4fab3d4812f431b53425a1fcf2ee2effdb4ef7e64685b83d7dfd3f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b1298d8d35d98ecab4089d44c51a741bfcb8e36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b97e03bfc710eb82285bccefee35e43a17586c30f7d57133e45fe5b28084de06 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..83d2702445873b02750e6dbbcdbea443b8688591 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897595_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2374541b0fe8265f5b4aa95a793cb70f4bb40ebb3e10b434ce658cd09c084b9 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897618_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897618_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..39263405c526f9517164bf464789986d197ec995 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897618_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:484827564128685f3f8b0d6b7eb6b2b78475bc5ee666cf690bb6121c6f177c9e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..208617f8bee8d5b3b7822d0583a099da0846c68a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4220db4427c371263b4039102bfe1cc8ceb9e244ab6e7d39d096e1427707a1a7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b083d412e536734d8291e1669fb5be9361eb0c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897664_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac50e48a5216cf4e90e9a7c7831f0115fffe9841f25c2e7fdc53209f744ab576 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897688_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897688_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d435d88e7a0a9a4336c1b8c127c39fce5900216 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0148_14897688_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e815ea4e2bb3d2ee4291ed8c9fc165a4d2f6b94400d24dc2765c34e123631eca +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879870_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879870_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..69c6a7b94fc671fa5036f220015ea255ea75c943 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879870_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4af332e93375b4bdea72251628d2916212e8701b9b814bac15e936b550720cde +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879878_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879878_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef40f6c340fdb6f3baf93121fb3e0bc7a529fc8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879878_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89af161d93aecb229af51960d3e4da6263e7c30545e48d73e387be9c2d5df147 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879892_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879892_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f84076ca1a739a130f10f824aca346f921b0fa26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0151_14879892_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58b2ec56a56b10f20dac019b4120edd48c2eb746c9ad7dbe9168a5b9e1704e6b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0152_14869815_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0152_14869815_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..7213c5c8013288163eb4486302426fb7f1badb6f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0152_14869815_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61da4618543b6d88d4f49dc6163cf9b9aac3ea92fab2d13375f4282a8b7a39e1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875579_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875579_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..dfb4e933eacb2c9473ac01bf2d4f58bfa507f12e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875579_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:768c0d824259f6bc7a1b352c3cd6221fcc1161d7085df184da28cce88b9d3435 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875591_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875591_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..86884c2a4a836674a4914efb06122df9ea000908 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875591_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:961c06e375e7ec6e218ad80f3c626bc2ed920fcce8f3edd1fa15f4004e990bb3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875655_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875655_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..70fca7cbbaa85ce60c0a413668bc29f6b4684ea5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0200_14875655_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f12d8b72caf8736419c34e8302bb25486b652c8be313df67baa3eac84ab69be7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0231_14869776_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0231_14869776_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..daa08f589097a34971cd6e10fdd0d4a7388df148 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0231_14869776_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:784223d4eb2eb750c62a9c9f6bdb5604b45a99e90c451e5e8dffa4b19db0bfb0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..d13ce32fb10d16b40f77d03d2c7d515ea9caa746 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ac2f1ef0b97b591b7268e9f729e8f0a9d78bd1e14ae41c0853f18745b0296a5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..179ab5a6b1a2d64ee0b169fcbbf5c4a2571883f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979210_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80c552b919a00c7f4ce01f6fa461306f2a3b84ea357de5626df2769b0a0d7925 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b0143d60c4b23adce93e8dfbc92265c2f98d47a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e3a86676cf23d3439ba3bd88d46ecbdcb38def0bd9ad971d244c42e2050c13e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1d83536999e073c058922afb4be07c3c7257cd02 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0285_14979215_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9afad8a3341073f34cd144628d6673f905f67e2a59522e8f8ce2c3212e126436 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14878997_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14878997_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..3bde95163ecd670d9671c3ff7d7a95b6b8a598f7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14878997_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60acf2f828c874b1a6cfef41e6189b72842f4718c286293ad62b87c32620cc8b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..42f59e0417e28bb8ef6fb733abc8d7336bd26d8b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a51a311191a92aa14d1afaf1bbe5b7a5562987e881bb16b16bc1da1f604411f1 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..46d071295173ff89a7f962a4af8708ddc3dfa809 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14880859_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba5ba99b45658d65a3fee18c98d23c167edadab788b691bce1b0e541e180d9f5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894049_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894049_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..454139f4a3c0e24657971585cf786918b23b992c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894049_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3034aa90a852a17d926ef981c0cf7191cc0a756827d8d69b007f53d186a9baea +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894190_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894190_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..f63f56966ec1aef9933796d194aaaf64ecff0317 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0293_14894190_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a395767ed5daf3e10c398bcf8dd3d4849bada2ea73990ba5febd3f5f39cc4e4d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..cecb8393db9698df795ba2b851b77723dff17577 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35915411d7f2e548450ccde6b54746b668b4b0b4ebdb9f8aa279f0588b8a8da5 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..50527fb4e67ed274d3385d7b9932090af6a8ed45 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875379_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9dab846c0eb23228445f84e7aab18b56fa14150858dbe542a8584d405b6a7866 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875405_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875405_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..00f4b17634e39280fc8d8f6851dee4dd119c29b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0318_14875405_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08211213069f2e260024f26eb76fbe6d0c6021bc5c923a19189d77bca6f81394 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0395_14867643_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0395_14867643_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..f580887f2e4345ba806f2ef13e151e8e31faddf0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0395_14867643_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0f5b73bea7406d6965d9b35d2dfb5f98fe5d745a672566d45032d631e750f6c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0426_14873037_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0426_14873037_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..9358e6627bd613b2f69af23fe66b3c692d5330c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0426_14873037_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21e6d8d730a895c1564a3ce5ce544e26a32075932999f04f2e57a4e702877289 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14869009_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14869009_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..00432ade487ad76c8e242b4eb3b1884baca70f29 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14869009_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a131d85455ba2a4a05d4340e2ac2aaa6ffd8c36fab77c7ab13edb9457b2af68b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..94531ee66c88d376a72a6348dac46ba7cba89bb1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4264e308d5c80956a13af6c6864ba8b28fc9e709a8420f426032c65bd5ca35f8 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..fcdf5dae915d20e544ac4353deb67b35d808508f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0433_14875693_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e90c313ab9e061cf2123a0c606d71ca1817b763ca03bd9229efd44220db3b52a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876215_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876215_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..79b034b98414cf9674bf721fc1f493946a3bbcec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876215_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e332dfb262ef8d8c2b8d235c20e99f1e396ad7118d374080095d5bd3b189dd7 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876220_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876220_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..91cf64ee6aaf1c6dc1372451ff8a35f5ea089c54 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876220_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bbf8b865d29b0592785474489710d28e4ad083778917ad27c46ac9b9fb344e43 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876229_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876229_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..425a0331c1253d0c6dbd79e9cd14b9a1a807ab6f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876229_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e729cdc0e9f260db1b2b2a50a83310e5a0f73485f9f539e4126e31ddb6674ade +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876238_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876238_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..9cc388ac698417efc2a8f9214641b365d8a60d9d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876238_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c31006a5418863cb632d92f3017b4d913a851533c20f2670f87b176d1257a3d +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bc5e00084f1bf873a050a216dc575bb517a6fb8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a872c582fc5a33776fb3302f78f7087ce14999704305cbfb026a9761ea06f7b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d8d98643fb0f903a6630f0a5ad772c30ec4c2e2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876255_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbd96a98e31009716f775e71806dcd4f026d27c55fdd5a02b3de7769cd6d7aeb +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876275_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876275_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6279a34aa6b3ce9ea16f47b2c9115a2d707cf9f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0435_14876275_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae66cfff19ad10fbafb63e158a4b0bf9ccce76276e24f1d04f79cb755f64f0db +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..4f98b49cb4cf1916d5a74d906a732281c881b777 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac29b75c1923e088d611f59b1ffd65e71ec8f777d49b49987fe682234d1af901 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..e504137c2c3c0c5a99577ac31547388e9655da1e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0468_14869824_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8134ccbe460d04b45f0523b7e9e23a00feb728429ed1a519763719418dba993f +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0485_14866784_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0485_14866784_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..781635632060163d2143d1a27abda44d0d8c9c82 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0485_14866784_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb27869523af38284c9b4492ebee3c7de3e4d2abe5782d4d74adf5b9d4112e48 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..25c5288a1aa93e106b46dd7399aed95faea1413e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fea0f13dd6213f9b4bc63cd534de47acda3802c6d057d020e14e40327884ae53 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..e6c48a7124bd3a4080863829c093f4310243c731 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0491_14869748_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20ec92fac41f1b49184bbb90330a080dfb74e42188ec4b8a73933d32af7f898e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0507_14869096_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0507_14869096_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..2507cdefd54b7a2fa4e541c5f16e43c8caf64c81 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0507_14869096_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc8c152356a8a9eed0190da4d8dececc97c01e0333a6be9bfe2b0cb72fe406cd +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980295_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980295_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..6f059d05ee6c2f5e7ed214a44957f42c35f70c8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980295_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dd48e090706be1b4737105458ca0e48eac3e3543bfc050047fa80e3b6d65198 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..2748adfdf2d91e66a9bbefc2aa249e5a42d5e09e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73e00059198cb0a62ee838f23c0b40cbf4e55c5297494b35824c2379258b8510 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee9c19d3b44ce0481d4eb00662f11342fcac4b58 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0528_14980298_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26d390e5c7f8a8c8519f99cc32b0ca47b4737dfebacc5cbd8da4b45bd67b6bee +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0557_14872046_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0557_14872046_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..3da53886b353fdd1308f1c993572c544c34c7c8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0557_14872046_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b733216041beea26f922b131bf9a72a0d0a6a6b934f1dd0ecd9a661fff5e248e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868883_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868883_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee92d0830337436cca85102040a1ce0f2fb43118 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868883_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:135f47fc014988de6e31dbcf3046cdef7ad1b5f40a7f08b8f6141c8a5e6fdc2b +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868894_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868894_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1ef9d4da5e4101f65c51cd422c3705a9eb0f8a17 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868894_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5359041fe7d81e0d84b20c1de21e32be24cdb01af2ea5294249ec75dd2ef491 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868917_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868917_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..163722c7c9f5afd48d2b474f70e290d87c38a388 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0591_14868917_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d96c288149734306514ea773cd8b8e3ba97ed559b633b5746e3eb21f435af7a4 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0597_14867507_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0597_14867507_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..2d17f723fb830a42376a8a1fb4b65d593e3a012d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0597_14867507_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d03acc7ac8c0edb77db40ebd0d9c3e208364768e399c564dd914f57b966c03a +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981549_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981549_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..379852a894065a09bca2acac6f477f0d5e879127 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981549_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02e0348a5d1868e5d2fec815c3350c8edb70df5b137f5bedb8ddd55eb379d775 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981759_004.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981759_004.wav new file mode 100644 index 0000000000000000000000000000000000000000..5a165aff3afba45437ebce3d1b9cae11ffd73e72 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0635_14981759_004.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14dbfacaa4666e9090bd36b4e7645ed16b1f5013fc39c82090cde76c7fc362fe +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..08eb91dc4fa36327b5b50bc4a3842b8f8da9bf43 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea6960a5fc546939ddba78404206a9dadecbcca77bff265a075f01c7c96f6082 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..9b8cc2b8598d5c39677e55ba47314682edcf8b8c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872850_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52d05d730efbd413cfad6b05479fc255fcd310e4e61f9564cc6290cc3559a59c +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872907_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872907_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd6422c77760ff13a9f6328bade45fd752510026 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0694_14872907_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95c07a38ad5ab4aaa8827ff2798c91eb36dfcd866f94349ec581c31189f24a3 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0749_14981956_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0749_14981956_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f96c519c97c8e6e50218c6b7cb799608bc768a1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0749_14981956_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fccb7ca84a07f7b87d5fd2a2955b42c2474e5159176cef54238dec80802aae2 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879526_003.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879526_003.wav new file mode 100644 index 0000000000000000000000000000000000000000..c608b724e40f3431d1e4aec1736c5b4f80037018 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879526_003.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58b11038f1e92c2d349859a5922fae9e20330d49d679ae0269d357f58a39afb0 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879547_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879547_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7920e8ae6f72f6a89dd3467c5997e91cc26ceca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0791_14879547_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69818fa9a6c6ca90320c7939be0753d06c6b59811f16915d6019fbcf2e5cc16e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979597_005.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979597_005.wav new file mode 100644 index 0000000000000000000000000000000000000000..4bc36a1d27a00a89ae9e09e78d54d2aebf5ca7bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979597_005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23549b0c26909aacff0f7566e0cc2fe0dc3d2f7a6e3c3f6451638242195d8400 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979606_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979606_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca85df50d85f117f9cc8ce7c343e2fe05e88639d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0799_14979606_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09289ebadc3efe344c9c020fa5214b29161c9b68c63af559e6cf1e17eed2fefe +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0812_14890365_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0812_14890365_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca5d42bf01318a8e23b9627ca4b57cbe37369c40 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0812_14890365_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e1e7c0bb4f1965c25fa9964bb3705150bf399e3e0fa0aa32a51641c128786aa +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892867_001.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892867_001.wav new file mode 100644 index 0000000000000000000000000000000000000000..efa9c849eb6ad027d28d58f1426dc57320cd2380 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892867_001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:943dce26ee662a563da504c43140360ce1b349e4b4d69b5cebeb573f7b7f5f1e +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_000.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_000.wav new file mode 100644 index 0000000000000000000000000000000000000000..62b93aa002d12ee1cdb5b9bec8d1561d0db25e16 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_000.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6365e138e022b566da8b0952339325bd11a8eb715df045af5a35258eebb06c09 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_002.wav b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_002.wav new file mode 100644 index 0000000000000000000000000000000000000000..1a6030e48b71fa4ddb8d2bf15013147ed1c0d746 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_CochlScene/Subway_user0820_14892881_002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f378fa630d2b762570bc9d913937ca161f2b8b709f1a067e9bb2f835cea0ef03 +size 882044 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1.wav new file mode 100644 index 0000000000000000000000000000000000000000..3740e7f1a40da3b3dfb4c00e66db0b516bbfc480 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4de81b6d7de07afd10dfb66dfdd93a5dacc6ab3cf3574831e6005a85417153e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/10.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/10.wav new file mode 100644 index 0000000000000000000000000000000000000000..7ef0fcc8d05eee91cf649f244d80a27c3b55f834 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/10.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccf49506df07fad23813624ffcfcead40cdec1f8264d80ce422e81bc3ed1036a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1001.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1001.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea844d148246e959a06712db8930b0712b17706c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1001.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bf2cfd51c06b1474a68a183bdc2e93d7df48f31726b9aeae3ab390c7f66f3e9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1002.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1002.wav new file mode 100644 index 0000000000000000000000000000000000000000..d8470c9867aa29781fb9081f50db60a9a3cb4d6f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1002.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28aff41db4f6aad8f7b96dfd74d6ce8d730528b97df30ca704bdf90fd4ad50e4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1005.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1005.wav new file mode 100644 index 0000000000000000000000000000000000000000..c4bcaf6091589a711b469d349083012453f7572b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1005.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3736a4c11170f16e0d911332654052ce4b7de57d62182b037de8b2fd2c114ef +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1006.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1006.wav new file mode 100644 index 0000000000000000000000000000000000000000..efd374f36f452406deed6f8798f125f852b1828f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1006.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0514c6ffb875573cba5449b1fd0049fb2ee8cc72e12219a107475fc1354b7927 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1008.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1008.wav new file mode 100644 index 0000000000000000000000000000000000000000..f026a0f2f5a4a0ad8fb94aafc34d34bdf2d85654 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1008.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a6d4ce95f445cbe6696dec12170859f671499a0d3031d6922a3248b856ca9d2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1009.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1009.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd13290f7ec8287d72645bcf9d250e946526a92b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1009.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f4f71baaaadeec09c6e5b6a980ebfac1da3d318d64b3fdab1c6ebb9193580fc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/101.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/101.wav new file mode 100644 index 0000000000000000000000000000000000000000..6942babed9a3340d4698d340f75fc7163cd83419 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/101.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9b3d22057488979539e0b0eefc7017d35401e494eb9b693e28b0e0c8ff066ef +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1010.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1010.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1ce233410b79430ec9209b90253e6d23fea4a1e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1010.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03be966882650bdd23c4d5f71dcaeb873cf2d0ebe43866d9d028375bfe450767 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1013.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1013.wav new file mode 100644 index 0000000000000000000000000000000000000000..dd7497ece6829f56dabaafd3086a17b59d2c5f78 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1013.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21d45ebb343e4a2de25fba1001501e594bae7c05998bf427c0fd027f297a1342 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1014.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1014.wav new file mode 100644 index 0000000000000000000000000000000000000000..6912a50501fab9c91de2a92bcd3f05fc117edf26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1014.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a42eba784d1e5a2e9f0920f1e6ba9a629728d88b1d89c467b482555f8f167660 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1015.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1015.wav new file mode 100644 index 0000000000000000000000000000000000000000..ae99391526d3ef50199127a05cdee25baff90811 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1015.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d2fd2643e394e1779167f73bf407dff16d785a0eecbf9c1fbc52efc89a36c02 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1016.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1016.wav new file mode 100644 index 0000000000000000000000000000000000000000..645b639635a5ce7136699f1ebc797357ddb512ed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1016.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8c300833f83dea2e8ae57992d4aa0baf7446f372323106703d4932c1b75fc5e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1019.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1019.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b9e0141d1864979ad97c4d7e293d909dcb684cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1019.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:807c483a9916a7cfc9ab0ebd2e45ee39d542f5cd2b9e58562c0d684f98707ed8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/102.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/102.wav new file mode 100644 index 0000000000000000000000000000000000000000..b5f2b30bad6b818e568c6187d2c0246b18101cef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/102.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4a88f5decb9eff5c72cf8539b6176b1e921b40730778af488c881b121ef4705 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1020.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1020.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0631fa114409f085ebd5772072668f072167808 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1020.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:853904cf275e28430add7d3715e798ea116838e2ebb5160b1ee250915f8de245 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1021.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1021.wav new file mode 100644 index 0000000000000000000000000000000000000000..95eb09488057b4cccfa049b63f84dfbdf1f79a5e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1021.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8daca72dfaa40a1d0ecdbdd324dea79b3a682d34132f3b7b07208e7bcdacb9a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1023.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1023.wav new file mode 100644 index 0000000000000000000000000000000000000000..34bd9437a69b264b4ce9e00275b09f8abfc078ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1023.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94b0431eb30c2e4714ee0e8d3f681a30353646b1599aae7817a7895f0f2b435d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1024.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1024.wav new file mode 100644 index 0000000000000000000000000000000000000000..3cfd4bc87d38a36c992781fc060c815630924773 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1024.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efbe9c1df53c52a8004ef5732aa2ad7bbdb86b32861e8d700e05d94dad664650 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1025.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1025.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb1ef890487369087bcf3411fcf37645d026a113 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1025.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8b69fc7d5ae1a31e0fbc1632c18fd80fd34bfa14bd764f93887905cbe873cff +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1027.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1027.wav new file mode 100644 index 0000000000000000000000000000000000000000..182cb499b009c84f4dabf88646b3d772798140e5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1027.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96587e54496b949a5a4bc4299e2ae1b228ed7364c8bd0550b7fd895cae77dd14 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1028.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1028.wav new file mode 100644 index 0000000000000000000000000000000000000000..2fc6f023919f743fc5ab716c40779e5650cce3ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1028.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29b21c0c36efed01667d2e5e5a1994903a55b4543f3300128d501eabf1fa0ed6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1029.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1029.wav new file mode 100644 index 0000000000000000000000000000000000000000..064e20eb926445abe234f3d9114fcf294f32781b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1029.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61d4fb55d82e9062745bdb59c913ed2fe6fa546ca8051c61924c8bad8c8e22f8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/103.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/103.wav new file mode 100644 index 0000000000000000000000000000000000000000..2336697a9f4c36c4cd773c5ef0715b965eef4206 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/103.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87c06ddc00dd15f6eb92340b622826e238a142ad1ea43bf416290252e940b5c4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1033.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1033.wav new file mode 100644 index 0000000000000000000000000000000000000000..8db0b2c8288951859e421b2db630f69475d114c1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1033.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d83de7087df4c09bc33400d44b461283b8c0da8ade3a09c20345e783f06eb6df +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1034.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1034.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6492a542c20d61534c6ce8c4bc550599503b692 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1034.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f20b437d895fa22876eb38bc64b763e0c2256c8db2c45dbfc8644c521de9a9a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1035.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1035.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6d3e230b90becf9c002829627d3fb45d1859e9f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1035.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45387f4aafbb9b0ef6f464137c8beb0c2bc863dad3b01b0c839ff21441a9c4f0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1038.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1038.wav new file mode 100644 index 0000000000000000000000000000000000000000..102558b762a72c6e24614459cf17fdec38f6e3c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1038.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a7590387a1761fd64305b9db011a4324f3ce652bc943c6a691f46a247ab5968 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1039.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1039.wav new file mode 100644 index 0000000000000000000000000000000000000000..888154d5e347815cd4ecd83efcd4708883553654 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1039.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6eb883798ab2d4cdf9d9dc5b2f8cf4acefd392e9f69cba041e5fe54b693c7383 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/104.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/104.wav new file mode 100644 index 0000000000000000000000000000000000000000..874e81f3ce1ff2c54171a0f02b59285cf3e6087e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/104.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed9344680b009e107231393b9960d884e69fddf63e59ad772fce4d2885d1db84 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1040.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1040.wav new file mode 100644 index 0000000000000000000000000000000000000000..1c948e20ca8f04703729d70ddf73eaa7a5e192d7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1040.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec599f7ee606360015a670255845b604725b8083f2052d6360198874051b571f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1043.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1043.wav new file mode 100644 index 0000000000000000000000000000000000000000..7376a1f09cc482aa037441abae6a027bc047a1f1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1043.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e4a433205d935eb9d292e56d3de263f763f0b812e77e864045e7cbb184d3d54 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1044.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1044.wav new file mode 100644 index 0000000000000000000000000000000000000000..af9244cb4c6f94e3bbd8c6d77205d99ec99499b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1044.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcd91119066663acb3648fa3d4bfa9a2114a8b582c54a73fe30e860026b5faee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1045.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1045.wav new file mode 100644 index 0000000000000000000000000000000000000000..6dc63f9f1794319e4ea43241c1d7ced9d0d9aedc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1045.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac6cdb47cb9a665bbe9cca998563b09a923173f297a69f4d993f52d32abb9b47 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1046.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1046.wav new file mode 100644 index 0000000000000000000000000000000000000000..a2e53c9ceb60b06caa69eceb00383f87082f4906 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1046.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5c880dab9c4db29994067ca1baa7441ab9e947afd361ab79dcbe3d9968e923f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1047.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1047.wav new file mode 100644 index 0000000000000000000000000000000000000000..6bc068a72bfe30d11e0c045703ed4749e6fcb052 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1047.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:749fc7395701b01758cbb80099031e1a5da73e70f59d20b3280968056299f1cf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1050.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1050.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e5384a0440d3d8f50815ce790507598c693c410 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1050.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8612492315292f827d74a49b9260d6d3884ba3da1b2188dce516d4a4bd3541e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1051.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1051.wav new file mode 100644 index 0000000000000000000000000000000000000000..e9922f2dc185a20fc04de38e679426d0407a77db --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1051.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7815ec2fc6ee76b826562a81a981cb7f5d2351a1113c59c4ab9d15d64a116e60 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1052.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1052.wav new file mode 100644 index 0000000000000000000000000000000000000000..e56c09fa45605b2ed496b70ad29f0ddbb4b92b4d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1052.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e38f243d566d57c2cb788fb976daea7d9aab38e62a760a39fbb8c5f94c1441e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1053.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1053.wav new file mode 100644 index 0000000000000000000000000000000000000000..37afa48547a6d8ff1033741ed8af19b8a58a7f8d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1053.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9ebe84ea4a353fbca907d7559ec31317db7197dfa92783b01507d51db85db2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1054.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1054.wav new file mode 100644 index 0000000000000000000000000000000000000000..e9d37944d794217a6af7b056831f6e0930a84ed9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1054.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c83f49b1118cc35167b9e96a511f39e027aff7a85dd79d242a0f655e0d9cfd08 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1058.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1058.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7b9b7f13eae093efd72692655d6fa5539cafcf1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1058.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fc6d39a2dacb20508fecef2bceaa91d1fecfb472d78d1fff40e92caea025cfc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1059.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1059.wav new file mode 100644 index 0000000000000000000000000000000000000000..24fe4513374f85ffc1f8778b71e1e02e87bf98ba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1059.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91a640f9cd28ca4a6883c61bec7a7a94679779bec400dda37934aa6530d32b3a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/106.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/106.wav new file mode 100644 index 0000000000000000000000000000000000000000..70e02ace43273d986b8e6d8724a103b8dea79a48 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/106.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:023767188a640dc18de650ef57057a16a67d16c44dc9051dceb5b4574690afa1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1060.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1060.wav new file mode 100644 index 0000000000000000000000000000000000000000..a86569561978bcc0c230637fc153824dddc38c7d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1060.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36ea8beb8b42be0b15ab768db8158eb8cdfec78f61668b6503fcf6532a515efc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1061.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1061.wav new file mode 100644 index 0000000000000000000000000000000000000000..b029892187192de5764c72f85787a283badfb49f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1061.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:383519bcf2e97afe6a7349c7f69d2362767027827015c89c0ed8a2d6c66e850a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1063.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1063.wav new file mode 100644 index 0000000000000000000000000000000000000000..e90c59114ab167065eb6a4bbd1a1a2e920f90c3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1063.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef3dd28ba2d9b4924f632002faba34f058c49eb3b8154f7dd2535b0b15f3c591 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1065.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1065.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6878732272ea132c90d382868dfd87cc9383a08 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1065.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf8e0376dd024e2c6a4fa9d8aaac2cef5d66236efb1cc597472a38e168287957 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1066.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1066.wav new file mode 100644 index 0000000000000000000000000000000000000000..c325e0159651895e85564390c7448a5fa868407a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1066.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab74530523dc2f742828abdde36f9484ae59fbd7b787a4374ef0ee6cf2b4d53a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1067.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1067.wav new file mode 100644 index 0000000000000000000000000000000000000000..0585285b751dc76a5f482a41ab94c0f5dcbe38ac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1067.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0bffd3490badc90d42641e0ee8789b8b13ce889667258f929ea9a6eed6639bb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/107.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/107.wav new file mode 100644 index 0000000000000000000000000000000000000000..d848a6828554a7f54e38c50cd7dd23f17eb080de --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/107.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b34e4a8e77a206893f9672331862a7cfb91b0356f4039a3b447e52c018622a36 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1071.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1071.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7a6d6b611b1a54234f652bd164fa3a35a616524 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1071.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d09a51b62f7545c9a03444b53a53b590f4689fc47a0492f2f038f6ce308c52f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1074.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1074.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b55aca62de377cfdb16779b699ccdda6b62b8cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1074.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33ebf0e1f57929e201abd3f18970bfc6ce283b1eb3a08b0f09d5ba43495990de +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1075.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1075.wav new file mode 100644 index 0000000000000000000000000000000000000000..545a005b021fba4b46638db30cdc1eecfa5a26d4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1075.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0248c49e4ff7a3d86040fcaf55763a6862b6adfc03bcca56381ca900bfe35dc7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1077.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1077.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ca55d832a7b1b228ed66c46c598f0e0e8416647 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1077.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d39ab3202eb53ec501e91f5036a3dca6e37375bf60a631753d7eb784b234fcdb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1078.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1078.wav new file mode 100644 index 0000000000000000000000000000000000000000..7a8aa2cf0b1b891a04384f73527d6a70d5a93dfe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1078.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b9ff476b6a1b51a90cb61fa239a3e38a28a47911822d6ad142c07196cd9f104 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1079.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1079.wav new file mode 100644 index 0000000000000000000000000000000000000000..54a1f7c266855cec2815ab8fb59a781444d96701 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1079.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a27bf679bc5814bf849f18c1fc3b8b2b9c32d1d229cb9d805351f3cb6688cde +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/108.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/108.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7b8abc9f7814cdd8043aef8c2c7eb83985607ed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/108.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:655fd790d002a40883737b7f83c08771861710c764f1d881e23d766a510078ed +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1080.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1080.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fc89ab1efcfbf0cb2e9f0817f84028cc0cfc08d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1080.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f938d53978917497209be1c46842cf8f54e8925811ce44800b0c481cca6f4665 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1081.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1081.wav new file mode 100644 index 0000000000000000000000000000000000000000..bc4c9382ac7241f48cefa10f748203cdfd9e1fdb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1081.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb1f8e434109ef7c995a6552b38a4e60ae543c002407d9bcae019492e7a46ccb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1082.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1082.wav new file mode 100644 index 0000000000000000000000000000000000000000..1533a10c032f8a68ce4bf12be1b015465f4d9aea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1082.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43d847f2764e321b993890a5788d51f57358d6a7abc4f5e7999e492914f43f16 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1085.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1085.wav new file mode 100644 index 0000000000000000000000000000000000000000..cc0b8decd2e81e3be722af07711dd7c081474c5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1085.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f59e9562fa60ca606c28e66f41ef03235ccbb69a1139a300fed53766ede805f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1086.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1086.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c8031ec0670c28a4c9f16f8a0cc5615109031d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1086.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98a3b2e993614a2a0fc7ed025bcfdaea08e571d899ca5894e62b886c34eec6b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1087.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1087.wav new file mode 100644 index 0000000000000000000000000000000000000000..1050251122f6bb7b7f50794b3163b72dc238d8e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1087.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:979dc8fe41cfa4e79f308c16c5959dfeb4fd81264f934fd797dea1f78f2e25c3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/109.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/109.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1f126009303ab8cfd2589fe2ccc4e8863047f36 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/109.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58dafcd02c2b12812f67db197008b555b93297d07c2cdd03c2877aebc57d1ac0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1090.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1090.wav new file mode 100644 index 0000000000000000000000000000000000000000..2be174dd079f7b55808b0792b8fd98c161e202f6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1090.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:813f57546c8c63638107f12caff149fda84ce8e39c8dcc817e9d0fbe8b434280 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1093.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1093.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b46fb0ebfd15fb671f4d61074f80b1a4405878a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1093.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3204d26952da857b5e2dfc728ae8af4b0af5fc8a515738af51671f2e4fc395a8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1094.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1094.wav new file mode 100644 index 0000000000000000000000000000000000000000..dd05eafb7f31ed227adf1a88194584d40f99a0d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1094.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53de53a387f29d9ee8b59e7f2f0b6ef8e0f2399a3fa4fc55d9dc57810e07cb77 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1095.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1095.wav new file mode 100644 index 0000000000000000000000000000000000000000..8728e531bf84d6b30fc78a1155985c9da462dfeb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1095.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e482c3db46967a264b4a48c7cbbdbb450b73bf86e387d5933173ce49bedd18b3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1097.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1097.wav new file mode 100644 index 0000000000000000000000000000000000000000..6a88096ab49bccc7c522ec74bea6d1e0c766ccf2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1097.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2db46dbd37f3526c77ca8c5d7a91a31e329226507f96484a6acdcce45234d9f7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1098.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1098.wav new file mode 100644 index 0000000000000000000000000000000000000000..05302ab620af0d48967260b6ec079085f43ac6f7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1098.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9107d9984d8e0c8ed140aa86c2a180be7aa8edcca55303d426f29c0057180e4f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/11.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/11.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc64513eba18829f4eba999f47a2689a61dce5bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/11.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bccf451cf81801a6e343996cd24789e8c7ed54bb41af3ac904e577bbbce4e29 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1101.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1101.wav new file mode 100644 index 0000000000000000000000000000000000000000..d39ccd4066c060f62262f51875c19b3ffceab61e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1101.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f711455e22da61383c20f3cd541805564861f237ca732647822b84db86de4a98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1102.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1102.wav new file mode 100644 index 0000000000000000000000000000000000000000..0a542333b21f21ed49cbe575556b5833eefda51e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1102.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b0de6fd86379a4acf9bc0bd8aea6c5342aeb489afa79609cdf44aa6b1c3e924 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1103.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1103.wav new file mode 100644 index 0000000000000000000000000000000000000000..537e341caa3c872e19a4ea1971807c9981c9685c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1103.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e449c61fa4fdcd65fd6a70b241cb3b1923bdfa0626926bb0702a04e289e3efe7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1104.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1104.wav new file mode 100644 index 0000000000000000000000000000000000000000..6ca8719dc94e2a05895fdca7ef76cd4bd05c8549 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1104.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa17ed24b728738a63e9a93cdbab2abeeea05599ece2f5bcd1741b86268336c9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1105.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1105.wav new file mode 100644 index 0000000000000000000000000000000000000000..29c76874b6845af749b4c36a148795c7197cfb5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1105.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce20ad1627790f0c84ba07992ea6799529b7e63201640a97cc644f8eaf86e208 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1106.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1106.wav new file mode 100644 index 0000000000000000000000000000000000000000..eea8fac07f0ee97c2cb2266ec42f4b7a00424163 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1106.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c616268941bd5ec79146773aabb50481d918abbad7ae4838a7fbce8dcadcfdd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1107.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1107.wav new file mode 100644 index 0000000000000000000000000000000000000000..9700067a39cf2d50735bf23a0bfecb63604c500d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1107.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:825fb70c2d9b6e5622472419698d124c75e66d1aa128027aaa0ac8d8577cb1c7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1109.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1109.wav new file mode 100644 index 0000000000000000000000000000000000000000..21fb626fccb67e64924db2c9c8dcb9bf71f93464 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1109.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5222092a55fc861e36fc667625ddedf4ba451118459525ce1580eab469cd9802 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/111.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/111.wav new file mode 100644 index 0000000000000000000000000000000000000000..ade8ea854ff9ae1589eb5a3879827234fa6626ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/111.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29b09ecb1dadd33f5daefee7188e0e601b528f2792355f98228918ea7c9aac0c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1110.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1110.wav new file mode 100644 index 0000000000000000000000000000000000000000..b022a3260c65416cc61f0aa62694f6130bd2c3d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1110.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15422b86751377b239c3b7dbe53ea85a40d5021697ad6ce18bb8f56b598772dc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1112.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1112.wav new file mode 100644 index 0000000000000000000000000000000000000000..628e3e831e46eafa852ad0d0b937797cb7f2dc55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1112.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e6d40d7c97c967ad174705240441fd27110be75e107a19246bac0c1e076c8c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1113.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1113.wav new file mode 100644 index 0000000000000000000000000000000000000000..c706b5ac07a6ac87ecbbd8e686f4b68e5e601968 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1113.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e210ca6567a71b20bff452ef99cc9c76bffce067c5cdd0dbb48bae829115e8fc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1114.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1114.wav new file mode 100644 index 0000000000000000000000000000000000000000..25ec9ab356f3b719f5c3a369c0f407612bda3fcb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1114.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72c8e257c37239b22076c58ccf9f7e577238022b57b0e961b32e6fd0a33986c3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1115.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1115.wav new file mode 100644 index 0000000000000000000000000000000000000000..18cad1c489e296618d4ed9e6479cc91672ca961a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1115.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27202f1ca678fb84adb74048d6114cc13efe1b878a1a5282e86fd1c43f788c01 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1116.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1116.wav new file mode 100644 index 0000000000000000000000000000000000000000..445b61dd589ec6bf8fcd1d8145763c1294303e79 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1116.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7efeba7ecd57003434063018c5ad9bfcf2e020717209b171c659866260cdd3da +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1117.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1117.wav new file mode 100644 index 0000000000000000000000000000000000000000..a273965f6116d2a550acf3cb08583a5326aa55ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1117.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60efb570660a471610af8334d424bf7e797fa687edb8302a445dd8f8552cf8cb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1119.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1119.wav new file mode 100644 index 0000000000000000000000000000000000000000..85238668ecb5dd45dc1d2cac178079ea99e5fde1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1119.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f60a72e639a5ede7b69b982f6cbd42944bc4fc9aeee828a664d1cec4950dafb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/112.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/112.wav new file mode 100644 index 0000000000000000000000000000000000000000..0d56cb213755e2149a0f69f24aa7acb427c82c7b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/112.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04ac28c6113324d24ec8ad04c2fe38416935d8407c52cffcdb24cd827aa4efb7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1121.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1121.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6e37277eb7234a2103215441051fa668d51d623 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1121.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94eb269b1f1f3c3e763ff97c701712aba4977abb8d3d66366355392b074b0249 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1122.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1122.wav new file mode 100644 index 0000000000000000000000000000000000000000..cf62c8f2ed78eebad581885fb342e4c90aab5d33 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1122.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e105dbbe5a35a37c3da0190c3f0db5142e38401bb87cf7812615f1e2e344207 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1124.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1124.wav new file mode 100644 index 0000000000000000000000000000000000000000..86eb7b5daed1588c56da657dbc62073f5f3382ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1124.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4df152e56df93a0d343bc34014a1c1d80823ad7d16aebe32e25b6b102a7e59f8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1125.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1125.wav new file mode 100644 index 0000000000000000000000000000000000000000..52f06027e8b611f87bbdca7e149fff7285f81299 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1125.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dc5bc1cdeda8994914c562056990297978b8d9547f8aa0089010812bb046d32 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1126.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1126.wav new file mode 100644 index 0000000000000000000000000000000000000000..7b91f25590b787239dd7f7463b07b935f71102d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1126.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f9dd464b0f567ddfce8c6a6b034003f08c510e9127b0198e3aa0b65b2f5daaf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1127.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1127.wav new file mode 100644 index 0000000000000000000000000000000000000000..1fe81d37a665623c312ccca20bea3bc3bb19bc6e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1127.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50a893f8c24078525b6fa0b15b559949e2c18919c2fdfa93dffd8768717857ae +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1129.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1129.wav new file mode 100644 index 0000000000000000000000000000000000000000..148eea3df980a9a9f2a8f89253b2fcc7d6697951 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1129.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a7dd58c3b8610c3988aaced86b3408f4ddae643cf981a84af91e7fb439fd75b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1131.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1131.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e9cce619268bbd256b8162e908b5b5383db56a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1131.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:416409ff22e3e23dd927d6f33f395f6c46b54aab09a5f458831755cf04d4c0fb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1135.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1135.wav new file mode 100644 index 0000000000000000000000000000000000000000..5706e340d5f1c5bb4a371ef825e76f528e382893 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1135.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2595a9d7731caf82ccb6eab25458cc37e5733f8f7390d85bdf8625175786a4c2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1136.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1136.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c92dc61f7c3855baf8388a074b097529f4b3521 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1136.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9327f65d6c3bdb6ecdda3dee92acc019d02baa47428ffa62bebc28b31900e090 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1138.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1138.wav new file mode 100644 index 0000000000000000000000000000000000000000..c5c92633b2cbbe980db71ad21f824e8fcca7413d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1138.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dee638fc3a2454067dfd89ace9250f8b9bf93c7531f087bdc1c7468adc90263c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1139.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1139.wav new file mode 100644 index 0000000000000000000000000000000000000000..490282bac47f46b9ea2f28ebca72bfeecfdff29b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1139.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b0ca581501913a3ea72407cd735c49aff28858ff131769a653feae0d5e7954c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/114.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/114.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc3d48897b60f5bd202a012ddc65c7156d9b120b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/114.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1295d9741ac5f69f9df9ad96a67200a9fa84119c5a1a9950d51b467a21a1d41f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1142.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1142.wav new file mode 100644 index 0000000000000000000000000000000000000000..fabb5d88e41befd9d7d48db4ee43ad0c9fba5f40 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1142.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3ca9aeca72c212fe072290479dbc40fed4b561325ab9bffab73740922d66850 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1143.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1143.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7692247e04b9dd1dee8788b8cc269826d1053f8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1143.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5098c06d68c2b58f7d23083bb1ff9ac5877b714324b3a65e8136ca57ad422c2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1145.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1145.wav new file mode 100644 index 0000000000000000000000000000000000000000..8590e386d3120caee3c2b9a2836a97765c278605 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1145.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ee4c687f8f61201f7bb229651aca7dd36f435713cff6a0cc39dabc4b5d6f919 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1146.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1146.wav new file mode 100644 index 0000000000000000000000000000000000000000..38450f3f0286da1b6b079874c5f952457371e1bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1146.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1452e6f327545a20f1303179bdca758ea28b57d1646f67daf74ca785013049c9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1147.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1147.wav new file mode 100644 index 0000000000000000000000000000000000000000..f753a1054b3458d81349877caca158c83c0900df --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1147.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90239c25e0a0e72beb258727f61e61de7462ed937fd88dbf610ab3c79132b168 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1148.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1148.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f1cee558c4fabc0d2092911eca5d55f77cab5e5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1148.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac610a0ea75a4ff72f7d36b257feca7f3358ebb2bd80d17d0e275f9e803feef7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1149.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1149.wav new file mode 100644 index 0000000000000000000000000000000000000000..b5694db71d1f478af660a0b19ce91dfda97c6a8f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1149.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de93df73f6fdcd75afad5843ab9fa7131d0b2b24bb8ef2fe3150ebeae35069b1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/115.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/115.wav new file mode 100644 index 0000000000000000000000000000000000000000..ffcbebea15ba37a4b84dcced06bcaf233b0cd670 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/115.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2012a0f357d4cf3c421f4a012580e8dd8a66ceec5d605347212839ffcef934f2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1150.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1150.wav new file mode 100644 index 0000000000000000000000000000000000000000..006411b3bec72bb9cc206ae5f9203b7462e4c302 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1150.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d813be01fcddef87603c019585d0ccef9de7d8c01b46c63790c26f9176136ad8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1153.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1153.wav new file mode 100644 index 0000000000000000000000000000000000000000..a54e8a10ce66c4b23ae5a5f80e0be75d990ff2bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1153.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c3c99cc3f468b0de6e1f0deee387b9fbafbcc47c359f8525a1723919ae38d94 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1155.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1155.wav new file mode 100644 index 0000000000000000000000000000000000000000..5d8bbd0d794fcb269219ef669220d2c7f08a9511 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1155.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83319eaa637a182b5b6a6d02e3f2613106ab4415721b9ce6033c3de49eddade3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1157.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1157.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a6bdcc717be2647564aa90f1e4e6eaf065b3f6c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1157.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19a2459cdd27b547648a50b5b16baa2600beb51f4bb4c4d67b81b84202ef0e54 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1158.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1158.wav new file mode 100644 index 0000000000000000000000000000000000000000..be6d81b97ba075faa475316ff3a265db76b585e2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1158.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e07a13eb882e9d8ea07b17d623376d3cdbac5c4000578229f08a123c68c5f98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1159.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1159.wav new file mode 100644 index 0000000000000000000000000000000000000000..61c7e992ea39ff974f8217472460769329a2d7bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1159.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16f6bb1d78f39cb78dbbf426bb7327b3ada704f82de35fe4f4706f762a49930d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/116.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/116.wav new file mode 100644 index 0000000000000000000000000000000000000000..b86f5a149aa5adce597379f3dc3f0a16b6fc07dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/116.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad2c10747c5d9739a86c7bf89932380b77de5f5bae0e8ba656cfc7dbef3b56eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1160.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1160.wav new file mode 100644 index 0000000000000000000000000000000000000000..33b854901ed9097087fe4253c466440cf6aeb634 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1160.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1941da0d5d82e7c3e7b9484f576a5b748e2caad8b0e5f07b3104a19699a4848f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1161.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1161.wav new file mode 100644 index 0000000000000000000000000000000000000000..1dedb52fa614909c2459d0a74182ba75a4f38594 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1161.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afd79a57cc7e565a5b36fa6af1dcb7445de0eaa527db96b7c5bbdf932c5c64b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1162.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1162.wav new file mode 100644 index 0000000000000000000000000000000000000000..13855b0b2e01aa66b84528345808f34d5c6b8d29 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1162.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:945d5bf461d031c889e147fd5b8b1994aa10d484b484855d3cc47a43ace402d7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1163.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1163.wav new file mode 100644 index 0000000000000000000000000000000000000000..881ffaa3fc8d89bcfdc6feae20aa0b97fb63ac59 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1163.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1eff8a7d2e8d0d5e1501f9a626924bb98e6d06b880261db1ba132cd14c85cbd8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1165.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1165.wav new file mode 100644 index 0000000000000000000000000000000000000000..34ed5baf78c55e7f56d0a3bc4d64cb52667e1a23 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1165.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c14939fabb952567e299cbb783437976c65f442c74ff05b2b105d7c83361a57a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1167.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1167.wav new file mode 100644 index 0000000000000000000000000000000000000000..67b21d40f1187ee94fc9086b327bec8fc74b9420 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1167.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65df89dac3191cee1d8327529f40694ec108f7c5618d9fb55efe69232ae5fcf2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1168.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1168.wav new file mode 100644 index 0000000000000000000000000000000000000000..de1966802aa8481a060cdae45366dab4065c1d78 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1168.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2db17fd3665510954a76f43f64616b9b2e62150dbd05d3db0dfc05ad2fa46c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1169.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1169.wav new file mode 100644 index 0000000000000000000000000000000000000000..95c381e6122439577ec448144e02b56e82140456 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1169.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1433363e6005eb1633a3d40bc5cef2c85f1bd0504c17b85d0010bce5947c4b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1171.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1171.wav new file mode 100644 index 0000000000000000000000000000000000000000..1947a4d1e1b52d1dd8e884861726165fefa6d940 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1171.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4f7c1df906aa72683bbef955041fc29f4b3853dec67042664745b4c87c698dc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1174.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1174.wav new file mode 100644 index 0000000000000000000000000000000000000000..4c848bc3e200f6eb4dac90cf9510aa5b0188b4ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1174.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7400ba0f322f911f0496b8a1ecceca8e32bb1edc3ad5c69839fe69f9e313cd41 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1175.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1175.wav new file mode 100644 index 0000000000000000000000000000000000000000..bad3bffdb1f2f1b54d5beed8fec05fbf42799e62 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1175.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8929f9de1f9058bef25462afb5177695bf1d4f5b20f076165b1638b00f49bb04 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1176.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1176.wav new file mode 100644 index 0000000000000000000000000000000000000000..8dfe97cd9a32870c6de8cdc54b315fa27a316af4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1176.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e22646e76c08f0572495a3803e0693131712e7cc298f856a48fea745175aedaa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1177.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1177.wav new file mode 100644 index 0000000000000000000000000000000000000000..3e67f723898e09c9c612210bc7ec49a603f7ccfb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1177.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc90b389b5ade315c4070b176ecc4151f3fc9f7ff4fd6820acde54b1f8e2a7e2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1178.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1178.wav new file mode 100644 index 0000000000000000000000000000000000000000..d546cd3c487272a33fe3bdf9aee1d096ad0a33ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1178.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cccc835a8a3e951b4e85be289ec210205cdc1edaa5ec312e79290e918cf540f4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/118.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/118.wav new file mode 100644 index 0000000000000000000000000000000000000000..9312d753de60d64d12031a4c75ff222c029d1dae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/118.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:650fd6572f2787bfe1cca64c79a23cbe51aa0b408054f1ee7975babc7568ac4b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1181.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1181.wav new file mode 100644 index 0000000000000000000000000000000000000000..48708d19d71076dc5696f34c23befa200cbb1338 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1181.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9986090f1aaa1be08bd2e8f8c81762063567946760044854ebae7951b95a40e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1184.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1184.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d97da2f52d6b80f3bdf6750af6f88e4a036bc1d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1184.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:767c95c4275caee7e25389dea03903f2de4683f158df84d611969d3f53bc08e7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1185.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1185.wav new file mode 100644 index 0000000000000000000000000000000000000000..cbb8d0cff0513c0a9a138e84ea903e46672b6692 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1185.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:900eb4b8d2af619ef513a772f2808ea1458f750872a45f7366f9642a26771add +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1187.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1187.wav new file mode 100644 index 0000000000000000000000000000000000000000..738e24896a3d3065b89af9248fac996558814019 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1187.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0752648387b9bc0a89d18660ebf9ed27364ced9991011851251fc6c71930b603 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1188.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1188.wav new file mode 100644 index 0000000000000000000000000000000000000000..9464b7c54437c7e7583ba0662f5af449ba5ca67a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1188.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:606a5d55d7e9ff7e87375678e953dc2010ddaecc22073778f15c1d0b3a9e0367 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1189.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1189.wav new file mode 100644 index 0000000000000000000000000000000000000000..d10939426b5249f5cfe11cc79ae22ec3192cfbac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1189.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:952a2e8516f7dd548d2a4c9cde34ff76a043be47561f3da5fa3406a028552bac +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1190.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1190.wav new file mode 100644 index 0000000000000000000000000000000000000000..51c1e6ace9cf84458bfa6d24ef8bef119df46c4c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1190.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a5648fc4c41bbe00a7b6c5f8c687e3b5ff924c3812089639976471ee08f5732 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1192.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1192.wav new file mode 100644 index 0000000000000000000000000000000000000000..88a46f6c488086cbe2208d3ed1a0efa87e45cc41 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1192.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a4b0d2ab9e67ede60b45b1102ff546708174d1af869eeb9c847757f3ff3d14b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1193.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1193.wav new file mode 100644 index 0000000000000000000000000000000000000000..7e6250a8927e63c2f9c58c925388677241c2355e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1193.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72442307891f9cca1d1f29ac538a8f5addaadd0d798ac7246cf8a106f107b9ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1194.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1194.wav new file mode 100644 index 0000000000000000000000000000000000000000..5c80efd14e6df673041c7643725a029c8a4cd373 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1194.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a1e2d39a990c1991112565b50b54d605aba246784c42cb1b68545230aa4cdd9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1195.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1195.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1aa927b1b434c2995cb94097c2a6cb7b3f4450f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1195.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa88b3c80e1d66f6e5f25a6fd0affb91dc1a1f1cfe08de6e5a06b9bd11033526 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1198.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1198.wav new file mode 100644 index 0000000000000000000000000000000000000000..a299c974965791f81ad977bffa06e09f56558a8e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1198.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89ff43bbda3776ebfb0a2fc7a70f2fb27f7221ec6e20168c6be935c7fcac3c06 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/120.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/120.wav new file mode 100644 index 0000000000000000000000000000000000000000..0571181eb72882c3f7f4d7b2922131f519664ba9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/120.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97a334717a39abfb29176b820ce3dc6252c26c7274c613fe9de7f9493c80cfd3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1201.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1201.wav new file mode 100644 index 0000000000000000000000000000000000000000..bc1c5e753a608b562e605fbfeb2be271cb3d9934 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1201.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a0c82b0cdcd06445844270b5624f27c8a7918f6a441802804c8e992ed8d94a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1202.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1202.wav new file mode 100644 index 0000000000000000000000000000000000000000..9326c969bb6c85c81c81093b41ee57c8c79910a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1202.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0ac27fc600f5f0a223c652fbdedc09a21d09086825291ce2370c901933b624b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1203.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1203.wav new file mode 100644 index 0000000000000000000000000000000000000000..cd9ae15530a85b1d856842e24828dd4c225b6b6d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1203.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bd0522481a8b4b29e8db1d6d750930b2f876f8f97a580475ec9b64166250087 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1204.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1204.wav new file mode 100644 index 0000000000000000000000000000000000000000..907ccdee1ee05cab72e76344c1d2ad82c29f4d6c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1204.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c823af6573c52f2b2b0d08596c1f2393e896d864e09bf5d6a82544df2d52fedb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1207.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1207.wav new file mode 100644 index 0000000000000000000000000000000000000000..cdd7825607c5e4040094d5b9488ca39aac6a62c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1207.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3906b32eede489b1679dea2b3444d9f1561a8e0e912134e9614319181580ba6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1208.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1208.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e17cb45eac92d73c1a2d04680b094fc0f1f8740 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1208.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:558e666dbcc46393e76acea938c4d5a788a0a4ae96e697d42c53b3571ccda62c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1209.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1209.wav new file mode 100644 index 0000000000000000000000000000000000000000..32c069a21cfe3af503510c6e4ce17b05fb175c72 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1209.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea1530cd250c1ba12480479109e1722858cc777c59fb7e7046b75f1999bb57b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1211.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1211.wav new file mode 100644 index 0000000000000000000000000000000000000000..f939635b3513c4a84e56b5bce4693ef16879b946 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1211.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62427a93a8bad62cd577de3fd1377a2b8726819353f30931880cf44bc08ebf33 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1212.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1212.wav new file mode 100644 index 0000000000000000000000000000000000000000..959022b83a362484d868717fd4ed112f728fe7ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1212.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b84d83e5f65db597234894dc13c21b0fa7d3019f483d7a28efdda061708264c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1213.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1213.wav new file mode 100644 index 0000000000000000000000000000000000000000..90e8b70ef99507d3cd825c01319450086d019737 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1213.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d3ab6a59c7d0e2185c6aeb82aebb7307def20825de3273cc91abc0c98da5824 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1214.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1214.wav new file mode 100644 index 0000000000000000000000000000000000000000..33b5101b52c903c2f468aba030d51ca92285163e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1214.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e46a5b7ef222a37ee2c7aab1808d4cb9cdf8ecd6dd2aac46e46348c867cbe1ee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1215.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1215.wav new file mode 100644 index 0000000000000000000000000000000000000000..650f274cb46750e003ce4d8b55d7bf361710e489 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1215.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef2a2bef107186d470c030f0beb193d5495fb06f1c52be09c74c28c1f5c1d23e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1217.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1217.wav new file mode 100644 index 0000000000000000000000000000000000000000..b433f717c8c9f2c681c5f236da9e1976311a98c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1217.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02617426a82ecb8d1b5f62a21f39def90b742d515d9aed7157eebbe49b2c70ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1219.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1219.wav new file mode 100644 index 0000000000000000000000000000000000000000..170a51c46f4bffd4c234b00af1fa916c3850fedb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1219.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0828867fd9dd19d546757fb36b007f8b8bfc1f8246236fa36839ca1085e7cf8f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/122.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/122.wav new file mode 100644 index 0000000000000000000000000000000000000000..66a61bcf9de08c1763b81bcd7f78c94709fc4daa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/122.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14a6c09c79db2ee0ce298f0e0fcb842324f774000dc28d8d804f7ad39fb92a14 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1221.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1221.wav new file mode 100644 index 0000000000000000000000000000000000000000..4538d9d7c45bc14102f298b65c2c17e0e4cf4c42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1221.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec3698d947ca69e432065ea6b045274373339fcd89b6227af09b9b52f840f286 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1222.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1222.wav new file mode 100644 index 0000000000000000000000000000000000000000..27ba6f604bf4796fe26f9bdcbc870d29d356519a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1222.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef74dcfe9f63868e59060aed036d4871ff1182a9092f020f98e0e7b7ae2fd352 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1223.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1223.wav new file mode 100644 index 0000000000000000000000000000000000000000..61b372d9373b0922142e8a595cde531eac6c2236 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1223.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d39e28e21a272e026d2bc7ccf1e8461e3634fed9c9ef061f333699a7f5aae01 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1225.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1225.wav new file mode 100644 index 0000000000000000000000000000000000000000..1dc7459344c2bc47e718e425ef03c2ebd74fe7f1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1225.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:651c4e9fd394792ed0a62f26ba56d02203cd40673749bea667cc67ed10539e83 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1226.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1226.wav new file mode 100644 index 0000000000000000000000000000000000000000..098d2f9646a452955435232a4d4f319a6908da26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1226.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f517801d49cc8eadba1bebee33c4c4133b4ee41c2e9b34b26dfc505660bf6eb4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1227.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1227.wav new file mode 100644 index 0000000000000000000000000000000000000000..227db667410e020daee2a47a4c6c773d574409c3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1227.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:473518448c088fcf69173e8d4d331137e0e5a4509f636f5c0c72340cdd583a91 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1228.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1228.wav new file mode 100644 index 0000000000000000000000000000000000000000..8279211dea7bf9dc9005dd549196f4993a961f26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1228.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00ee1fd645f6fb98a8d00b420d732269d16b7f4bbcd9b503471355319292555e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1230.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1230.wav new file mode 100644 index 0000000000000000000000000000000000000000..3ab64be9561ef9c3aba80d4e7d41d34e3fdc91cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1230.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:858a41b6c9d1fa9234636ef69710e8392ac3e111a9c279de08ad6027ce0d762f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1231.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1231.wav new file mode 100644 index 0000000000000000000000000000000000000000..e7a13578fe28b08925153ce6ff3b1ca00f4cd3f4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1231.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b9fa02f5275739cfcaa6e934a2880ee28e1bf00e54501b80615d1390d2c1e51 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1232.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1232.wav new file mode 100644 index 0000000000000000000000000000000000000000..164858b9f97befde73f192486626bf50efdcdee6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1232.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b2557566d6b77461d9b2a82a4ccc70e18e50b3378156d0f632c9cec876ecc96 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1234.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1234.wav new file mode 100644 index 0000000000000000000000000000000000000000..672f21fd06483ad790ddfee6e337a7f4eb332a53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1234.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83db0b650a5681f86258d22ef48d89aee2f7f408088dbea77c93fa1740f1ccdb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1235.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1235.wav new file mode 100644 index 0000000000000000000000000000000000000000..167df5191a2df17143d51d974d4487158748f221 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1235.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:349fb1666f2a5fb8fc7b83163c3aee768ce4a51bb9003a5bb4cace44afc28411 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1236.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1236.wav new file mode 100644 index 0000000000000000000000000000000000000000..6e234b015ba3e39eddcaf99b0f238148d88e499d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1236.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4765b4feb24ac42ff20a2b274fac1f1ad48d0a85363f1b31df4dbcf818bdf5af +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1237.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1237.wav new file mode 100644 index 0000000000000000000000000000000000000000..27199d11a8c4618f0bf98acc780b46387f17f722 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1237.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:907b70107fcf445a3bcbd54aabd3df5da08529c38ef7ae08e1cf4440fd458933 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1238.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1238.wav new file mode 100644 index 0000000000000000000000000000000000000000..9180fbe9bc1845502a9d11543a0b6d65f08cbd7f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1238.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04a40c05fc23cef01a49d6ea20212712faf1b4e983f197de9bdbd8d61354fc3b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/124.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/124.wav new file mode 100644 index 0000000000000000000000000000000000000000..231aec0fdf5d098b5a1e89c82481ea90ab812703 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/124.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6716030825587ef4464df1d405e7b20191e6d2d04d0d3d0449b7ca4c6fcbab90 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1241.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1241.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b5a39c8210c61786c9fcb96e97744e74b03fca7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1241.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee82423a603e2db5230a7592653c953b7e6a2e4c2e32d2e5896425fff5982917 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1243.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1243.wav new file mode 100644 index 0000000000000000000000000000000000000000..c5fadf4a104ee9136e3dfbd0737572c72c3d84f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1243.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbea04059fa2fe2280f5cd6969c1aa3063d5f9d22f5e54cfd9507dd9b35c99a3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1245.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1245.wav new file mode 100644 index 0000000000000000000000000000000000000000..a04ee1c467bccf459e6cc987dba5987cda68a157 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1245.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71e0049d958d46f73121f2905a7847a8fec6e5ed10b964d56b1238525d977113 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1246.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1246.wav new file mode 100644 index 0000000000000000000000000000000000000000..d5875a3165bce0e2b0e275bc1d205e747b42002d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1246.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e9cc26988246c4d5d60e5f54e2afc82e16fd546c9e2b70aa8161ce1e22091a8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1248.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1248.wav new file mode 100644 index 0000000000000000000000000000000000000000..eaccb985e71080b27e4272531a6ebdc7080bd6c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1248.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10e181f8b16455edb73023fab858945710779e0ae5de35df6a71b4b7bfa1fc0c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/125.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/125.wav new file mode 100644 index 0000000000000000000000000000000000000000..abdc4d626bddf5c3218fa08501550fd8f2b96cff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/125.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca188a336f5a9798ee4fdbda7f57842b03954b1d8ae9915c1551bdd30bbe363b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1251.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1251.wav new file mode 100644 index 0000000000000000000000000000000000000000..f887d1b4d88843c7075e8176c1d566ef5b60e566 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1251.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a4a21ed2becf6b14364689dd3a796ce2563f8ad245f64940b869a8efc6c8c7f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1252.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1252.wav new file mode 100644 index 0000000000000000000000000000000000000000..73d74c861b3352054fba5d46ef60b3b838dd6723 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1252.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e44505cc89d2ee3824de6e4e3c2661f7f4eab3a240c6ed124648bae7536b105c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1255.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1255.wav new file mode 100644 index 0000000000000000000000000000000000000000..df26ec2d2138b177e15ccf11a5a8bfc478409c70 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1255.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4e370208ee4ff934f11846e9120c0d7a8219bbd83f67321722cf9e500ab3b75 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1256.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1256.wav new file mode 100644 index 0000000000000000000000000000000000000000..70e66e594b7f7e84126a10c04be0d8a3779d0071 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1256.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42178bcc2d64f57869073a66e103da9f701453d1e48fbc625ad965a39d1ad018 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1257.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1257.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0342d4fe7a5bf8be78e9af14779294e64d24f40 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1257.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:337545639971bdee5d82431caadbc61eca7882121819f852bc176b26f92f708a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1258.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1258.wav new file mode 100644 index 0000000000000000000000000000000000000000..1890fd37837a654b4c8408cd24a68b70c2808da7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1258.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da28482b7089bb270893da98727e4c07ba42ab1f57f8c229cfb5393eddec186b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/126.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/126.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea33a89aea98454b7a9e2b170a051ba730e58d37 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/126.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3d6ffb3445c1f3c99f8d327882e8a54147f80e7382862561689522849ca23c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1260.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1260.wav new file mode 100644 index 0000000000000000000000000000000000000000..12559862ceeb6394b53852c1b298b69fe17e3f60 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1260.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b99564377f9b32b73bb1ac361405ef1e91306618a3076e3ee9249f2938df2d10 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1261.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1261.wav new file mode 100644 index 0000000000000000000000000000000000000000..19ff3a4b9e18b14a47ce72a3e09fd4db0a95de5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1261.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:752ed03cfb91ce6eef89e746fa218b22f20e2706177cd1d5ab7c902173741436 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1262.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1262.wav new file mode 100644 index 0000000000000000000000000000000000000000..622f7079c35764829eccbd6a8607b8be002cd2ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1262.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c6e85d0a5c8d699bbe3d6bc3b42a861d742bf003c24827f900aedba7a78420d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1266.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1266.wav new file mode 100644 index 0000000000000000000000000000000000000000..0fa13dd87f9b18bdee8954070a027ae0b0f97913 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1266.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9a2d57cf8400ad16ba6f0b270fc175831af7504b00fbd9e13c0236aeb93eb2c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1267.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1267.wav new file mode 100644 index 0000000000000000000000000000000000000000..e2dc871735f77b42df613827aaddb0370aae47f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1267.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5b30d51eed6d5dc4b1b0ca9120e45d258c196c0e3dad8676b8681d0f95c01b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1268.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1268.wav new file mode 100644 index 0000000000000000000000000000000000000000..424510545bbb6a7cafdb1735367789adb431af75 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1268.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a02807f2659b4f7d9c754ccae5a98d5a1c2e9cf9f714201230b6c4cc1e245fdb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1269.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1269.wav new file mode 100644 index 0000000000000000000000000000000000000000..1eb5a46a798f02255dd5dd18f48aeb87048d9ca8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1269.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:741865105624f4cc1da4c9f1ff740cefd78858cb01b19db63cf24cc2bce97501 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/127.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/127.wav new file mode 100644 index 0000000000000000000000000000000000000000..68d36365d4dc855857edeecf08cf667420cdf681 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/127.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e8d7a6b9f276b0d193eb41e8d60a5c5f6828efb67d813c5f4dc74a732d2e672 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1271.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1271.wav new file mode 100644 index 0000000000000000000000000000000000000000..eb9de05ce01e72708a0d17bfd07264992f853f48 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1271.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf63b7b612fd5828ba4624a00b30208bf2043cf3f3394a6df91bcc8b5fa92d10 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1272.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1272.wav new file mode 100644 index 0000000000000000000000000000000000000000..872b2c8f5a6ab39868951acab988c2939e2dcb11 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1272.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18bfbdc37b929186c149544c35d0576982e4ce7d293d27d708e0542e5d451a36 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1274.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1274.wav new file mode 100644 index 0000000000000000000000000000000000000000..67bacc395449348c8d03695399f0205bd6e3f7ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1274.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a44b194ae6855fb88a37ed0af6e105480dc2dc3d07ccf02c174dc14ee7fae962 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1275.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1275.wav new file mode 100644 index 0000000000000000000000000000000000000000..4427e9b0a4df7378fa2108053d9ef60897608484 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1275.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b926b5718456ce5aa517e74ab281dc0578189d33123fbd8c412daa3a9b32710 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1276.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1276.wav new file mode 100644 index 0000000000000000000000000000000000000000..669cd92ea5d0cd2964ef35dd5bb70bd6f24d0124 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1276.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a71d7d6eca5e76de94fc1e1b70a4cf06aa6f0aebedef5ee41b1d8127c83acd33 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1277.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1277.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdd9c5453c2d69ee244d152563e577e1cdd4a5e8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1277.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0881d7df28eb53c9f68dfdc905b939dddf8d2fe8d6a4391509d8f4e3be0e055 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/128.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/128.wav new file mode 100644 index 0000000000000000000000000000000000000000..77f8bbc96bdd249630ed337b7044858a782488d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/128.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8e8bfc85fb95ed2442b4667697e24d11ffdedfd33127c565107ef489ad11f68 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1280.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1280.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b2bad56deebd29d75171da563967f8f86622312 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1280.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f406db75f356c9d9253b283a47baf97f0da875696c573139e0cf9c574a64ccf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1281.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1281.wav new file mode 100644 index 0000000000000000000000000000000000000000..155d20cfd2b142b9a6689c5ec1978069fd0e5e3b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1281.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4eb09d68c969c8991a4f89a246e49573cd079a5fabe2e94ea2dd8a24d4df9bd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1282.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1282.wav new file mode 100644 index 0000000000000000000000000000000000000000..c2cc860b71076d2eec2d42340b5f88f37a528251 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1282.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ee744443811fba853c0507084d2a836f1e669599baaef1cccfed4593ca983a9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1285.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1285.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0343619facbeeb387fb3095a82f50041301bcfe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1285.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:232a9f1a70f1f53de6e1f04b00093f401b1ebcaf1231895d76ca82af86e975d6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1286.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1286.wav new file mode 100644 index 0000000000000000000000000000000000000000..473bb738ea5d20fbe6a440d4bdc01d73d6fcf710 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1286.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9ac0fbeb266039175b800690586b3446f006694521db11523c167f34fdfe909 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1287.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1287.wav new file mode 100644 index 0000000000000000000000000000000000000000..0c3235e870ca69fa328301a382b3ba5b1f33064d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1287.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6ad17039e3bf3fa018ec545e2d8aea2589f5cb1a03b37998c1a7a9039c6b7ee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1289.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1289.wav new file mode 100644 index 0000000000000000000000000000000000000000..4edd1155d01ae8bdc117f5ff30a5e610e468b494 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1289.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3714cf4e16151c399abf1b85e280cf807466d3fde63d3dd3bd9d0a1d49421a93 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1291.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1291.wav new file mode 100644 index 0000000000000000000000000000000000000000..e2059391ffce7adfa721fd8e3281dd5130abc4d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1291.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce74a60c604e1e30fe4051d1c7d11b6720e65944b5b061dfa48aab19684ae4c6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1292.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1292.wav new file mode 100644 index 0000000000000000000000000000000000000000..3de7e93b8f209379a4c40c887b40f9911ea31ede --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1292.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8faa54829ece249bb9e81df833d55147fc85019fd97a0a146d061833eb2c59db +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1293.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1293.wav new file mode 100644 index 0000000000000000000000000000000000000000..c131b1b334f615fff0ed6e85017319ba3ac94249 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1293.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2c10a5257f93fa78007802e95afecabb26d243c735a34d90ee353c4961ad504 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1294.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1294.wav new file mode 100644 index 0000000000000000000000000000000000000000..bcc57b3cffd2e86b864583a0c552f0e55cb6d12a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1294.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fec439655ab38cb9dd189d2b8c4311d6bf5e8f74aa07373725210fbc4699a2b1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1295.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1295.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca98b84c97e97be690397fc0aafda566cdd36700 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1295.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58676ed478b52768677a3179fba3816bb67611c3b426d6cd9b0673e0e999a817 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1296.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1296.wav new file mode 100644 index 0000000000000000000000000000000000000000..64e1dffccd37b2931a02419633a39f49b530963f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1296.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea77f5cb0bfb5860d76543a6389a7096ea739e2ac87dd544719b30e9123ca736 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1297.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1297.wav new file mode 100644 index 0000000000000000000000000000000000000000..f81a929aa762b1153645870a20109a6d4bed01da --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1297.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cede0ccf91d31caaea23b2a860231e8fb71a27b0d3bc703ce3ccf9ec54a8cd6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1298.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1298.wav new file mode 100644 index 0000000000000000000000000000000000000000..52086017a7a5285c187ff40e712c66e47d148b8b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1298.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da61f680430fb3e1cb569ee233461b6983a7c142efdfdc60515ac211740d3855 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1299.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1299.wav new file mode 100644 index 0000000000000000000000000000000000000000..837084203dd57591a68b1c40d1bdec7bfbced09f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1299.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:502be82d346ca28df21af801ed63270b1bd335cd58ff2acd141e28ea317a0895 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/130.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/130.wav new file mode 100644 index 0000000000000000000000000000000000000000..21299fb75151bea9daadb8e25b44ff3e5dfea672 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/130.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa2a17ac6943bbc35c497fa21e30458667de1c4743a88a8eb115105d2b3dc886 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1300.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1300.wav new file mode 100644 index 0000000000000000000000000000000000000000..3148079e2863e65df0a0ed7ed93f38cc96c2d2cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1300.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efd5db4a155f3d4dd2107d5fa7ec75caa84adc1e0588c01ec629e18d161e66ef +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1301.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1301.wav new file mode 100644 index 0000000000000000000000000000000000000000..794565e18347e57518d877b3d16fa30f3d8f8766 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1301.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff10b6fead2df2e021db4fa2f0cf6175757fef7948980c930ad4a212c14c9ab6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1302.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1302.wav new file mode 100644 index 0000000000000000000000000000000000000000..48de01e1b53aa22de2088e918748ca9dfc55324b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1302.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7481a144694c4f333c3f466202e9e870b8a666f09c72894add4c6b482543f9d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1303.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1303.wav new file mode 100644 index 0000000000000000000000000000000000000000..f21ff732af3601e52efb80db9a308b6bd6f63f06 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1303.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b81ad13786f232e4cdf4d320ef9afc913a7b026a3646bf0c7424d20ac786b23e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1304.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1304.wav new file mode 100644 index 0000000000000000000000000000000000000000..440676fdf70cfbae9e322e92b18b88e16c8f031a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1304.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f11b99e6b997e6001bfb8b7f1469ed07166b9a79c23927c39a0a4f6fed328f0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1305.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1305.wav new file mode 100644 index 0000000000000000000000000000000000000000..4af8665411ce6e51454a7b051862af11fd893bc6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1305.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41dcd6dd7f8edbeb1783f7698976ddb0e101fec4b232be1b310229f05bd0e6de +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1306.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1306.wav new file mode 100644 index 0000000000000000000000000000000000000000..bc422af8578bd5a1631286908c2ba4627d74265d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1306.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8246c1947a383ddd1881d0e2c99c0db820ec44f1d934ed26c91f11ef98492f2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1308.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1308.wav new file mode 100644 index 0000000000000000000000000000000000000000..b9e94295b3bcb51a2e52d2457173cf9f72dfa9f4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1308.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37b654fc4d8836f045103a97f655b447f1df873d31303e39f10e6afeec1b54cf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/131.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/131.wav new file mode 100644 index 0000000000000000000000000000000000000000..497f09473d4ddcc43e087284143d44001b051852 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/131.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4337dff08591740e03ecc9ebff04862c95e5aa4f2389be0d14a82a417971ef59 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1311.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1311.wav new file mode 100644 index 0000000000000000000000000000000000000000..641db0dae4e4238026964e4cc3b50d48f3168534 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1311.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:579d2a19726534454c7f626a7309ae3321af6288d62b09ff0022a95fac5850d0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1313.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1313.wav new file mode 100644 index 0000000000000000000000000000000000000000..d45003e4edcb4329057c77854ebb4e6ba51c282e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1313.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd54e9c149dd16ae76631377e7b06544c64061403f5747a72f2ef1ed96ef470f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1314.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1314.wav new file mode 100644 index 0000000000000000000000000000000000000000..6472bc935f42f10fd2a89ff5eb9a0b1e3f7cdb0f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1314.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9679279f72b894befc28eb414ab6242c0fb2c716e27077760c38cb0d71540b20 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1315.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1315.wav new file mode 100644 index 0000000000000000000000000000000000000000..3eb661c28e9b00dc4fcccbfcf344e4f7c0b3b9b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1315.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5dbbc6677c28284e2e27ff32c5a860b39449d8f9880c4fc300459077c5cf02e3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1316.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1316.wav new file mode 100644 index 0000000000000000000000000000000000000000..4281676ffd717f4d6ce21eaaf1bf843942bc03d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1316.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d98725d27e929194bc48d2dd71b4c5ee135ebe39ab77f37f27cdf4b9086e7f99 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1317.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1317.wav new file mode 100644 index 0000000000000000000000000000000000000000..14adbe5a6f9dea3da2f59715a922a3c0cf2aab09 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1317.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b24487a4f85bb1f549c8d1ef954d92974db79e88574c209edb3680d011fd47c7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1318.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1318.wav new file mode 100644 index 0000000000000000000000000000000000000000..da6c7948e85313175fa1f5c52755ac9cfa01a51e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1318.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bee24829dba93b5977d40f2724195c3b8ee2e8e7c8eccbd1ffc73046b46c44c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1320.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1320.wav new file mode 100644 index 0000000000000000000000000000000000000000..c5c4f1d30c3f3edb6eb207c161c68f751f6019d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1320.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84cdba66baaa5f6e9d9cfa93089c2918dcafc4316f6281d1206ca1bd9fcb4d25 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1322.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1322.wav new file mode 100644 index 0000000000000000000000000000000000000000..24a12540f3bfa425cab39d6c99c3f02e2d31bfb3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1322.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f72359195921f59c0b7686d010d128e691d7a6ea0dc3083a5ae747c0c14f5bb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1323.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1323.wav new file mode 100644 index 0000000000000000000000000000000000000000..1769c1565179935f6e649ee81a13cca2cf7add0b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1323.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd1e4dd0b1daabc9ace48c084a1dd42f5d30caf1ef9b2b9639c8c7b697228edb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1325.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1325.wav new file mode 100644 index 0000000000000000000000000000000000000000..7ff232e11768ce07f15fc514fb2eb7fd9afb6e65 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1325.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e65e9726a28c4f1d7947f86df5e45e07b7a5a2222b701e432d9527a8b0a07c8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1327.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1327.wav new file mode 100644 index 0000000000000000000000000000000000000000..2461922ed03aa19e027c3d11e216a9705b74b60e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1327.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ac8f9a379003023aa9d5e8d06124147756d88bbd95e22e734269ecf9fcfa777 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1329.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1329.wav new file mode 100644 index 0000000000000000000000000000000000000000..82c492497f147dc335f5686b17340fc8e597fb2d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1329.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48e44ae3c5e47927c616d53cfdbfda292f9a58d36ee1317d42558d88df96febb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1331.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1331.wav new file mode 100644 index 0000000000000000000000000000000000000000..c502c03d39f43bc3b9a7566447f74cb22e459aa3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1331.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a20be46d2ad338f4ae11a719aded8ebc0114080ee2ea8178ac81153dfc375e11 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1332.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1332.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0821402261b48a1e102472de1a8e5ca61cc983b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1332.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7f7f24ff227a242f92f0a1e0278e3e49a2bf880b944e50d0e404a0cd8e83a29 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1337.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1337.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee107fb46812831dbbc843aa94cd0e757fc57687 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1337.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1757150132a3017566e72f4172e0a6f208e69f83ddb967ed5d27f65ef963d05b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1338.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1338.wav new file mode 100644 index 0000000000000000000000000000000000000000..cc70062522c6c37c1e886b4a6724aea3dfb2d538 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1338.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf601896e948b8e38b3f39d4b3503653211933b7d18869139bb2ee610b082125 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/134.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/134.wav new file mode 100644 index 0000000000000000000000000000000000000000..02fedd6e371ea7bdc43d607da23130d95fc4fa5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/134.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ef94aaebea3427141f81281a9157d00bf317987cd192b23ba02f6c639232ff6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1342.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1342.wav new file mode 100644 index 0000000000000000000000000000000000000000..34d1e8545d30ae0c2fcd749f2e9d4516a97077bd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1342.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec94a780a8d8f2dd2788fb3a6a67e6555be0659554dd7f1992183ccff048f8c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1343.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1343.wav new file mode 100644 index 0000000000000000000000000000000000000000..21595a3dcf329d3fa18caec118656ae5c0498edd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1343.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fe999e038f20d92fe1e84eb7a79a3c96af34c09a36084db0f718369a685e3c6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1344.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1344.wav new file mode 100644 index 0000000000000000000000000000000000000000..293d82139264e43500b61e75b9db1f15cf831322 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1344.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:595445a85f903ffcd5bb3d7af2b7d21d508f2c772072fdff068956dcb874ac42 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1345.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1345.wav new file mode 100644 index 0000000000000000000000000000000000000000..d9fa43ea9c15fdd05f04ff1c50bc9ae17eae4a95 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1345.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a6fa663d3c0be34add748317878a025dd18fa06f4ed88614cda572a7f27de89 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1346.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1346.wav new file mode 100644 index 0000000000000000000000000000000000000000..391a3206a19b36bd8249bbb18abe6ae1818023f1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1346.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:899895acc8483386d80ba4b9a40b2ab7928ca87ec5e53bba8bf267e1909d3f19 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1350.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1350.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2ccdfb158faf7ae5d62f7d3c918ce294f3c6cb1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1350.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:304dd6ebab17396254945a4ed564ef1fd33e51e3d60a915912abb50037c84af3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1351.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1351.wav new file mode 100644 index 0000000000000000000000000000000000000000..6143415ebaac7c8635b3cb546a2c3d84500e1c2c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1351.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67f5a5c9a149c74ae201477c879f8f206591a8c4738117890b100fce6fb6ec51 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1355.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1355.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7726d4d6d3d76a67b38c890183f4c5969b473bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1355.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0efd3f588c917c400bf8a5d68a29300638e48f1443c546da167996de6f7aa0c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/136.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/136.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb5a3329d4dfade6116898f6540924316b528229 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/136.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1baf89d2527e8bfdb72553cd00f66248caa46c4079eac76b75d87481924974ba +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1361.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1361.wav new file mode 100644 index 0000000000000000000000000000000000000000..40712454d44335271f8bdbce5ab072e3a11d9cba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1361.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e306d42293e6563bb569d2c084e193f364dc27a0c08e3ac9980e25b1474ddfa1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1362.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1362.wav new file mode 100644 index 0000000000000000000000000000000000000000..4ccb22a525c14e4e5a6ae95b7b9cd9f663085c33 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1362.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:941fa30e17c97d0bc7bc558e2a3161c0a4c66f19f6d8bf8cf3a54706f32f376e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1364.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1364.wav new file mode 100644 index 0000000000000000000000000000000000000000..e985b0b4efe9cec3212b2fef345011e8adcda3ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1364.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99eb1a8ad644fd4be7c3cd1deb15a77e93bceb17a7d4bce84aae487cd24b1b10 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1365.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1365.wav new file mode 100644 index 0000000000000000000000000000000000000000..35562f3306661c14502dea4498d6f7cb17b5709f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1365.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc690ff5cb6623849f9efd6823e2dc59ad4eb8145a2ce6d1dcacf3a682833a75 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1366.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1366.wav new file mode 100644 index 0000000000000000000000000000000000000000..0096dee96a2b3d53970714d3d7bbf4bfef55c163 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1366.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98b28d4eee06bd2d266f286a925e660024c388a368ac251f4ffe0b0230d66730 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1368.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1368.wav new file mode 100644 index 0000000000000000000000000000000000000000..28d5db4b25cb3c9315116e99fb14ea333b4785bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1368.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58fb462645a58177fe01e1940602e760cbff807f501f36121895e2ca0a2f3001 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1370.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1370.wav new file mode 100644 index 0000000000000000000000000000000000000000..2598d10d4c56612f32bbc135b27354ba667ec2f9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1370.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b8540aee589f5e9d14442d776e75dcfbb1305c492b5e3dd15be254c5b90711c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1371.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1371.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0d8e4649a4d4e956be53a1ca25565da994e4f7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1371.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd9564df4330927c01eafc83f7e6cff392926b730675e93f1c6201ea14740cc8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1374.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1374.wav new file mode 100644 index 0000000000000000000000000000000000000000..f0519d78916e6dbe6ba6a0871f7384664f4bf71d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1374.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5898e0bcd13a443b839e4bd6af084269053c06743b69dc5bf2286fdf0519e5fb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1376.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1376.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd468429d774347f1a0ba62fa23a0c3b0cfa3165 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1376.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88ee31026ffcdcafb9f2bf37582462226860a4892d5d1c14881e46378c059db8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1378.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1378.wav new file mode 100644 index 0000000000000000000000000000000000000000..4266027009b7311ee2cf2466fbdee5648ecdc09d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1378.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afd9b845fefff2a9c270300aa4b2ea996ecf7432ef27014a06def2b95579d520 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1380.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1380.wav new file mode 100644 index 0000000000000000000000000000000000000000..e79b8e5d010adc65e4600bcd6ccf9f93d851af24 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1380.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a4706fc9609ac7b4db220c26379ff12d3350e3d0f82c946bb421803d466d819 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1382.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1382.wav new file mode 100644 index 0000000000000000000000000000000000000000..94e6e9e33b8a2d966af034053fc842a1d716c923 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1382.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b25204f94a7bc42d73ac0fc93a2be295bb4a3a448a7e800cbd2e41609a90eb3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1384.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1384.wav new file mode 100644 index 0000000000000000000000000000000000000000..356331c69684c90994cd93c6002be98e9722e4aa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1384.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6688c598a89592bc2698561c11bf9444f759e8334c2ea7ced5496594195583b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1386.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1386.wav new file mode 100644 index 0000000000000000000000000000000000000000..4fcd75a687caf00794c60a50dced29363406f333 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1386.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7649b249a475fde15f36900a68d97985c4937649fb0a6a61a0b7464702479f3b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1387.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1387.wav new file mode 100644 index 0000000000000000000000000000000000000000..6c5a70c0d7584f50c5538512af76303a3f225077 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1387.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81307bdecb39fc243866d8a6f1d67f94043f0065d907555e6e778d344803396c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1388.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1388.wav new file mode 100644 index 0000000000000000000000000000000000000000..ec7c07666abceb5c025e2ec4a5347490b681b87b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1388.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:765a6181d45aaad931afc862b7901fee4ea3fc420877f7307a7eafd0abe2d1b2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/139.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/139.wav new file mode 100644 index 0000000000000000000000000000000000000000..7f8367f3a2e5be10dafdc3462b407e3090d5f25e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/139.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37a9516b466c872fdea20541e04f6d41a728639f09fa287697acd5b9c8fc2a40 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1391.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1391.wav new file mode 100644 index 0000000000000000000000000000000000000000..ddc8f6d373d44e47a02e6af3873e56279df63c05 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1391.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:960f23c2cc85c1ccaae4b0889d1f3aee940187c0de724b315cd193a0eed22e76 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1393.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1393.wav new file mode 100644 index 0000000000000000000000000000000000000000..1dc24b03ce97c24e91a09fdcc959350b6ff58905 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1393.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1701e9a0e9dbbe51bb8641e82c68d37cf65d6626bec2d19f86f2758363c46334 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1395.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1395.wav new file mode 100644 index 0000000000000000000000000000000000000000..20b463dada1709b141de0bb32fbca07990074d70 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1395.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7931f9a4cd5fe679534d2ddaabbe3e3b74161f6ae320a949f6343eb20d244d15 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1396.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1396.wav new file mode 100644 index 0000000000000000000000000000000000000000..d2fcdfabcca56fa4167fc7c832f64f097aadb1ba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1396.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc141be5448fe254c31d9b08492a50382077259e5e26d6e3a129f78aeb280cdd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1397.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1397.wav new file mode 100644 index 0000000000000000000000000000000000000000..12528aa1e5f56db8015045c90e2c64ec37895e5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1397.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8545b7a02f6eec647b18c033f09480c7099f197ab30c6c4cea9bb83582a11536 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1398.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1398.wav new file mode 100644 index 0000000000000000000000000000000000000000..2c370c76cfa9be3f94f7eadf91556e0f786d4cf4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1398.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60db613931868ab3862115f59142ef827dccb2177dd616d6b2fc7eab8a6c1d45 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/140.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/140.wav new file mode 100644 index 0000000000000000000000000000000000000000..016088588739db010f749fe60ff0aa0cde6795a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/140.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:489850ef8da802e4e568de946f5cc4cf7f818dd147569614d9c85ec89cee029d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1401.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1401.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1e1ae37e5f9109c7cd9649e51dac5c8da10e0fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1401.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b7c26a02396ea1973736aa8a549e9c1bb04cfe4fdca38dede16b91cf994a96c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1402.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1402.wav new file mode 100644 index 0000000000000000000000000000000000000000..20fec27d24879635303b87cf305a41657ed5b0f6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1402.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89c2c13b49601a04ff85d19884bfb6dba22b2982fdd71122ed6219d12ab2b748 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1403.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1403.wav new file mode 100644 index 0000000000000000000000000000000000000000..87cd600ab116fa736a73a66e949994266282b75f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1403.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84eac42280255a94ecc17cf1fba2dd326201489a7e2d028a9f7772e28aad4d58 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1404.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1404.wav new file mode 100644 index 0000000000000000000000000000000000000000..5be1e8e2d2167dd815fb9d9cc2ec5d17b95a80fe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1404.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f372b2c3cf9a2accb9908dda48d2a5aed19435327b80eb482883770b1beaa795 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1405.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1405.wav new file mode 100644 index 0000000000000000000000000000000000000000..632d5119a68c2bd19d110bc7aa1d9de9bd66f006 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1405.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23f4ad3f8bfe108a9375281049dd8a58a0491307c3214186e34eb491137612a0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1406.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1406.wav new file mode 100644 index 0000000000000000000000000000000000000000..23e48600fd43e4f0e42719268737628d10b66a5d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1406.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbd80ad83fe3164a994a311e908d051b07e6a8d043368aefe819e2ceb1b4f399 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1409.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1409.wav new file mode 100644 index 0000000000000000000000000000000000000000..6048b40cd1b742bffa0eacfe39b79e9f5c91b9ae --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1409.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f60f0880468ce24819b89adb53dee384caebf51469d3e39431b06df45b5fbf2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1410.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1410.wav new file mode 100644 index 0000000000000000000000000000000000000000..7139de036f3fc950580fb1676b57a28d21784ee3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1410.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4ba6d02c38be5285f6b66700be972a29aa1213413507e8550bfc66f497b1183 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1412.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1412.wav new file mode 100644 index 0000000000000000000000000000000000000000..06ead1d94d5f50d0c33f282181b5673d018092fc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1412.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebdc3a15b75aba12e2f6d2d2b45b03e0dd5cfa19802e4649ad232c39ab747a72 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1414.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1414.wav new file mode 100644 index 0000000000000000000000000000000000000000..f5a2ebd3e2bac09e9f15f43cee1739ec4108aff5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1414.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34b1cca02e9989fcda79e6efc0fdeedad40f05c81b6316be0890b7565ef2ea38 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1415.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1415.wav new file mode 100644 index 0000000000000000000000000000000000000000..69cb1671b9e258377353de75d9200f0c2a3008e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1415.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:151dada32a4299680918e65bee8c63f7a96d385c116f6933d318387848c9b761 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1416.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1416.wav new file mode 100644 index 0000000000000000000000000000000000000000..9dc55ad169bd5df4025631663d35e79c246ebfa0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1416.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60e1ae7fc63d4b945eba4aec4c8ad3bf29e6b8218a07fd9573b825874ab65821 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1417.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1417.wav new file mode 100644 index 0000000000000000000000000000000000000000..fa1766b512f3e8e8e123d8a8519865858d30339e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1417.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34ff0fe89fe0a3b408989d9113b3b62df177f9cb23a4da680d2e0fbb5b70784b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1418.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1418.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bbc079dfb4b8c4bd5d527b2ca73cb2328c102ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1418.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6eae58db6e5788612190aad101a7dff932eabbae5cbdd3aa95f4a87753624ce +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1419.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1419.wav new file mode 100644 index 0000000000000000000000000000000000000000..b14a3bfaf3578322b82f314506ff7f9a90c4339d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1419.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f2d430a0feda5e40e62602842efd001baab48ed5600d0a3692e3d507bc74e19 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/142.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/142.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f1e385ae10781c7b7175823bc7256c17647e967 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/142.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fd40929b40b40857d9072ac4ddf83b932fb424e1ddcb0bd9a3b2cb52277da10 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1420.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1420.wav new file mode 100644 index 0000000000000000000000000000000000000000..8ce1fe4c20b50367a2059118dc48d5610e0d35f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1420.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:741316a743a28be5ffbec55385e2317eafc97d0a69b6effbcefbbbf15973080c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1422.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1422.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d7f85713e15301aedd28dabb262215c5e35d485 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1422.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58911baf27a6b9f13b045154f374b9de50e6f1a698c274ca04445168238b0d91 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1423.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1423.wav new file mode 100644 index 0000000000000000000000000000000000000000..9d99b8a6e513118755c99d2a16db711e3550e42c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1423.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1701ed859be5fccfb0e35909f43d3f1b1480b61f53ba191925511bc26be5bbe2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1425.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1425.wav new file mode 100644 index 0000000000000000000000000000000000000000..a9d969c377ee612d37c5851f3dfb1ea5493b8acf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1425.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:939e0bc48a88b93b67d977f2b2ca1049baaca2bb9422d0ea92373d9794b4d68b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1426.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1426.wav new file mode 100644 index 0000000000000000000000000000000000000000..57e29440f9a3fba66653d27c5bcd4323f1644625 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1426.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1543f6d9519937f7e975a13985dedf3d94fa2cf0fb490192299b1b6364bd7799 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1427.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1427.wav new file mode 100644 index 0000000000000000000000000000000000000000..41c1434cc73941c7ceb1d5ae2ac4cecd391a602d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1427.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c6a64b41eba969f5a66bf5bf19813f3e75cac72728877d224342e7cbf050f3f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1428.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1428.wav new file mode 100644 index 0000000000000000000000000000000000000000..941270bfcf959a71a22338078f2fc30e90aafa68 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1428.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb54d81256913015491ca5f2268aa87b5d95904cbc7a5adb48a97a753f9fa59b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/143.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/143.wav new file mode 100644 index 0000000000000000000000000000000000000000..80d8a87046b9cf43d91c93de7d4570afd50e2e1f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/143.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea75fdf0f6235661613e114757acb0c75ef87fc3bfe63ca6ecbec60f4bf210c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1431.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1431.wav new file mode 100644 index 0000000000000000000000000000000000000000..bcae94ace08aa18720e13c84720fbad9714cd0bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1431.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2753f27640cf173879efeb0eb14ea76c203566abac2725831008f08574ab6757 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1432.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1432.wav new file mode 100644 index 0000000000000000000000000000000000000000..f362097ce4a018ef3b74e8b087ba6f0bb9e829cf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1432.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e8bb8b71b80edc5f98284349e5914e5ac3f5496e714a519284f5ed5665a29f6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1433.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1433.wav new file mode 100644 index 0000000000000000000000000000000000000000..e7af53894c9784889ae6cf4901bfc4419e593e59 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1433.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45c6b88061d0aebc47cf4e2d06541237d20197f253630968c118c9b62045a9d8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1434.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1434.wav new file mode 100644 index 0000000000000000000000000000000000000000..3cdef6ee0760b7853491ce1d0973f2a44a508a22 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1434.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7008fb8309871de4bf070d70a6a132e2e2f7975ba02be48cafa46961feb2e6f5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1435.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1435.wav new file mode 100644 index 0000000000000000000000000000000000000000..13b76a6da33bf0914c43a7a616fccbf2d623cc75 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1435.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9fd38933eae7a1f02cf2caf22779fc260d4b5bf7e7149df2a8d01a378460e25 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1436.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1436.wav new file mode 100644 index 0000000000000000000000000000000000000000..4e26472a221072c98c376fdb6696cb38aaa46406 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1436.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f1f546dfe73413f96e59b5aaab2aaf1a56e727c94469cecc03edbd1d074953f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1437.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1437.wav new file mode 100644 index 0000000000000000000000000000000000000000..953d8ba6e1d0833499043290bbd81d1f81d78bce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1437.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1cd13e368f0550ec8e634f8e0947bef77682043ecdc668c4c911651f665c68b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1438.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1438.wav new file mode 100644 index 0000000000000000000000000000000000000000..be0b2ea98657cf3d0b6032d72c5c5063011a4087 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1438.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de8c3e037ec43ac7c0258d6442766ec52ccac0aa1b7386efcfec25dd4baaa6a6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1439.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1439.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e2c77bc93d1a4d675cf0b36b1f60a5b455723c1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1439.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99ed3b0834931bc52fa89fd9828aa6bf97f6f1149510501afd6fa63edd2c73ef +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/144.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/144.wav new file mode 100644 index 0000000000000000000000000000000000000000..efeec2e004a0b161cb928d9f5c7d33163b59af1d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/144.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8163b042e23a7d81ed631619604648473744cf287083b0c3ba12d7eca0aa15f1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1440.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1440.wav new file mode 100644 index 0000000000000000000000000000000000000000..a8297f16f1b492df9bd9bccf0761e514f7f580a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1440.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68d41b924edf1f5f746b080268bc09a9996ccf51b7f72074d29252e6c1cedd16 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1441.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1441.wav new file mode 100644 index 0000000000000000000000000000000000000000..b75768c9231b8f517cbbe8efa95b86c92844a23b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1441.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d9a4ba2e8f195405effcaa248263702457f7f5780680e583eaf834b2b0c8f7e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1442.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1442.wav new file mode 100644 index 0000000000000000000000000000000000000000..07fe1d826411cc0c9b91e2e5ab1acb96f2ceed66 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1442.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fafaefe50fc8f61b1600d6e56cb87ac565fa6d621f00239cc641f0a61ca997bd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1444.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1444.wav new file mode 100644 index 0000000000000000000000000000000000000000..3b3663c748781c59be184cfc35282f5ffea2aff0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1444.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:733b1263e55bbe2bc37400e9aa0e29ef4001b20207ef746b18f699b073e742da +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1445.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1445.wav new file mode 100644 index 0000000000000000000000000000000000000000..f4a82485d73196e3d95418cbfb09e27b5a0e34ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1445.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06f4c4bdc60fe26d9b9781901041b58fefa9642d1a0ee3f086cf44546e35b6f7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1447.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1447.wav new file mode 100644 index 0000000000000000000000000000000000000000..261d96efe830aeb5b5b253701af72d5d4e4ebe77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1447.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:963859e06e9692a75ef9db17d260c89fec0650b3d901b11098d7513a61bf3316 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1454.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1454.wav new file mode 100644 index 0000000000000000000000000000000000000000..f6ce877e9cf4234c500136b4dd462b9abc08b2b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1454.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22284e2712e8bc2ca9db9b31775fc3c8e56d207b58a0a4d3ad68c97579e3b2ed +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1455.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1455.wav new file mode 100644 index 0000000000000000000000000000000000000000..6cde9e0e98ead56c78456546dba1b8abcd989a9d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1455.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec1f265d8745543f02f4ad81d69563c53339648b7e5239834712258a39d8eb76 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1456.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1456.wav new file mode 100644 index 0000000000000000000000000000000000000000..4c7643976a3d314b55b7ab82ff1496a3480ddcd6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1456.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d89cf49938e407b7cbfce2e0477aa6ce18f0757391e685ea181fb58495291d86 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1457.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1457.wav new file mode 100644 index 0000000000000000000000000000000000000000..38826d05e5ccf05567f41601a816df2ba21e96c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1457.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a912869bca73785907dbd4b1243356d8ff1eacfdb329da4dfd531e908eca009 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1458.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1458.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1ae0d3a6998d6cca5aea65c94f8d349bb86146b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1458.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6902925591f94bec1f4551f5c0992c69ffe2bc634bf074f7e99f530abee6256b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1459.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1459.wav new file mode 100644 index 0000000000000000000000000000000000000000..4aabfcd4021a8cdcb167f962365dca0f62841324 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1459.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e70b540f99220e9ad96e5a3401437f1ae7ffa3fcdc75a11f1050dfd9c0de4e6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/146.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/146.wav new file mode 100644 index 0000000000000000000000000000000000000000..824537eab201ab9c9e9e2ce932d6e1bb642405cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/146.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19fe3ba7e76b00f34469d4b12d57bc615cfa0052553e6fda33316bcb64bd3e1e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1460.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1460.wav new file mode 100644 index 0000000000000000000000000000000000000000..f1eb070deaf264c3ccd5ec6d3e6c9681f0108d84 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1460.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d7ddb954bb98484f204395450c5e7b937289051ec167e3a23d8c88d8d6f3a9b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1461.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1461.wav new file mode 100644 index 0000000000000000000000000000000000000000..b04b18cf421195f0ab74536ef0e310b9f247c008 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1461.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a7e4cb8d2f5b4f32ca621dda76af83419ac64123c45435ac9dfaf7ce5b1f2a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1462.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1462.wav new file mode 100644 index 0000000000000000000000000000000000000000..1be39c8442b980acb486a32132066ba4ff615dc2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1462.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:927b4c4aad359d3c3cf729f7bd3ae6e903f3d3850fa44358234a1d2b174a5f9b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1465.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1465.wav new file mode 100644 index 0000000000000000000000000000000000000000..15147ea4957a6b7275b7b9b79fd547f3ad7f9427 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1465.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca4bd0b6494fd4820b37fa94441d8d6b27fc6c94a0937f4de96b26479edd6280 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1467.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1467.wav new file mode 100644 index 0000000000000000000000000000000000000000..f19ec1fb9b4839a9c84c5e3c671871e51da26d99 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1467.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ec560eb1a941453b37339ac26c4221e1be0e17dbf7004b33d4f586639012ec3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1468.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1468.wav new file mode 100644 index 0000000000000000000000000000000000000000..9749057e02be263eaa23e7ac0a550c664dae02f8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1468.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43a59317c051630f1f11fbd79ea35898c069cba03dabdef5bf70614a8636e2b4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1472.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1472.wav new file mode 100644 index 0000000000000000000000000000000000000000..950f562c7da0439f56e2fe346dffd6f25f1fc338 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1472.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be088c23ed70612e0e282b281ecddc764a32f272235a73280377bd8e5875675e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1474.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1474.wav new file mode 100644 index 0000000000000000000000000000000000000000..a781b5c24c53b4d143316cf767044469a4fb1ae7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1474.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6670523bc1884f0e12d939e6e99c31f4470c9d32da4009da2f58eaa2c2e07894 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1475.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1475.wav new file mode 100644 index 0000000000000000000000000000000000000000..ce16e607fa68ba3ee8191a5f753cd659a17cfa14 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1475.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b91afda7fee590885b1ddeb7ca9647ea55bc36cf4247d54d92b51ae436aef242 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1476.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1476.wav new file mode 100644 index 0000000000000000000000000000000000000000..47b66ff96bee076aaf191436e1e8ba4528e477bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1476.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27af1ccc780c16d758b3a44a7adf418f8b8339ae8287b458faec39b4bc942a95 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1477.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1477.wav new file mode 100644 index 0000000000000000000000000000000000000000..88b0e63d15712d4eca98d4b96164afd942ab32c0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1477.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ba26aba7174699c7ae31cd43e764d6b1e7ea58e6cf964875abf4be8ff9b564c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1479.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1479.wav new file mode 100644 index 0000000000000000000000000000000000000000..654284d614cd3aa5acbc397d4ed7180ee7bcd46f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1479.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d2e009749863bc1b44277eb38dbbb07722cf115c0e01ea649bfcf293e5fd345 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/148.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/148.wav new file mode 100644 index 0000000000000000000000000000000000000000..ad0629e588b4f54631828962a3ed8c84fade682d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/148.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:771ac13e1084b3bbcae2597e6c89b0508a2545220bdc530a673b3089fa3159de +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1480.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1480.wav new file mode 100644 index 0000000000000000000000000000000000000000..89de506af82ab49d6669aee51b1e5d260f8c47c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1480.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e5710b6eeb272c424b5a3614d449595f9006942217479dd8673b4a4f6bc0f05 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1483.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1483.wav new file mode 100644 index 0000000000000000000000000000000000000000..2276aa64fe53c81bc392f78c92a371bb2c926d5d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1483.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:549ef51e5abd59635d60d56740c511e4c5ac550d0f99387ed9e6eb55d2de5574 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1484.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1484.wav new file mode 100644 index 0000000000000000000000000000000000000000..d3e22422b1975791874df7110df0ec7e37acc2db --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1484.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5a0cb502417a82e733937b93efd0fee73a590e521861d201fc2ab3491cbfc22 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1486.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1486.wav new file mode 100644 index 0000000000000000000000000000000000000000..8e6cdb7a7753cd8bc661316f1c577c8101544835 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1486.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c038abf2d216c503e013207f3ab8d35e7ec3b36582fa5637f3680d8e8059994 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1487.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1487.wav new file mode 100644 index 0000000000000000000000000000000000000000..80081725164b7a2ea596bfaaf26800c2e467e308 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1487.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9cf12409b3cb3ea5d6ab0214cae31e95d172e1ec06ccbfe6867650973b9acd2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1488.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1488.wav new file mode 100644 index 0000000000000000000000000000000000000000..22d922051d96b3731e368060e7d68601374ca2e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1488.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6609b7f89ce5b72582807ce824f675de00907681024a354eb8adcdde26830c34 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1489.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1489.wav new file mode 100644 index 0000000000000000000000000000000000000000..936fb5d10311701efeddc9cb7b220c5f2ea06ad3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1489.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ff4fe4fe81623095e6b632abede54ca15e51ef968cad8a8f4b432e6dd04cbe0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/149.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/149.wav new file mode 100644 index 0000000000000000000000000000000000000000..30473c0b014112a2bbe8346d99efae849c83e00b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/149.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ec61bfd795f97f859b2632d622878cff25b58865a2104a7bb41751468a667d6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1491.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1491.wav new file mode 100644 index 0000000000000000000000000000000000000000..75e02184312a71f4d59c43f115d5ed7db5c10cc5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1491.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f00490a53794a183e089eeb761f5b96d922f6c36465670cc91b1e50f89667789 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1493.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1493.wav new file mode 100644 index 0000000000000000000000000000000000000000..c35d71948b1b6c4e092ea47345821aa47499e1a5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1493.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c5908d56e449f9929dcc84d2accb6cfc0b23c200524f41ce6b84e063461245 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1494.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1494.wav new file mode 100644 index 0000000000000000000000000000000000000000..0c4ab39d6ed8bf07bcda710853e647b7d7b17a34 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1494.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a6b94150e29e4aa333b5cb876422e8cd874efbc7c51394ad8a342c0d167cefc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1497.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1497.wav new file mode 100644 index 0000000000000000000000000000000000000000..772a39891237fe7f77859203626b102ded3495ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1497.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e47c51f013c5c7f47521b9296e7685a170e9a9917dbe965dbb82d504e365ead +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1498.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1498.wav new file mode 100644 index 0000000000000000000000000000000000000000..4673519ec6753e4e1c33134115d210774605f823 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1498.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f884d52d4fde7ac54183711fcdd3adf7ca18eccc4c076bc03c443cabb60c2618 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/150.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/150.wav new file mode 100644 index 0000000000000000000000000000000000000000..08812aa27dfe2f9da51970003eb4447fe9a8d907 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/150.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28a643c9f694a6c23f5d51a95513d7f77b5cff7208ccf3e2c6fdaf9d59f85765 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1502.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1502.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0fbe910291b1921ee9238ea6ca2fe1fecde255d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1502.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4033ed671422d14d9cf4f383fa8c43d5593c4ff9565e3f7f5bd40bf062ffeb39 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1503.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1503.wav new file mode 100644 index 0000000000000000000000000000000000000000..927452623f6d06d11559b184fd4efc8031ba06f8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1503.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d9c0b7455fcf847eaa25f27a77508fcd6cb60745985ec83895ba898847505ba +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1509.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1509.wav new file mode 100644 index 0000000000000000000000000000000000000000..92f7b701cc085991136d7ef20b501166801e1c2b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1509.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:044dc9102d892561b7f5fd7fbb9ddd8d66337f841a30043245133ba97ac9cacc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1510.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1510.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c1288db04fef5669884b475c094b29264dfd94c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1510.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3901ac19b4f55d1888f9ccc74c4cbeb5440fd12ca02d02d04ff52e225710016c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1512.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1512.wav new file mode 100644 index 0000000000000000000000000000000000000000..f421b4de21ddcbc50fb1449e34124c336852805f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1512.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4608573f5be94ac3de82fd45b7259b12340b94bf9d0ec0b309c7253ebc42beb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1513.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1513.wav new file mode 100644 index 0000000000000000000000000000000000000000..10454312def36c0a1dd58d4e9d46e9986e4a3241 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1513.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ade65b73249fe4b4b79ec126ca1eb2f4f54082e6be82bbea63758a986129371 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1514.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1514.wav new file mode 100644 index 0000000000000000000000000000000000000000..79d8e5e6acd5aafd2e008d4363097bc86fbf0c22 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1514.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90c6bd1bc50379bd37dee6ad33e0e67e9e9d6fde2e95682be692beb0415911fe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1516.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1516.wav new file mode 100644 index 0000000000000000000000000000000000000000..8afc8b9ae8f086d452f203015b945091b64e3cca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1516.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32f4406f6167c3a2a891a52b808cd0f6bbcebe58a00b4737b74186e57cba4808 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1517.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1517.wav new file mode 100644 index 0000000000000000000000000000000000000000..440af14c1f2bd217ebb6742d603599053e3ac573 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1517.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c88c048f73ca533fdcac14568bddbae4c52ded9033d995ea4e4100871b978d0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1519.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1519.wav new file mode 100644 index 0000000000000000000000000000000000000000..17df98709121ffbe5368f27a37fd3ff2556fbd1a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1519.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ae62eeee7e004445535d6a2820c1810057851b303b8fd5b81f98554bd323d26 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1520.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1520.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1d6059c6c61e244431bd60e196a144b312cdc27 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1520.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2467786314b88c2297c976cc57b0af990a74ce92e8945c6bc589b5c7a063d3dd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1523.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1523.wav new file mode 100644 index 0000000000000000000000000000000000000000..68c438396540e95903bb3489407b4150d5da867f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1523.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d16226877e753b1c24a5337604cea0f10345db1157836c4001f2546a90d07400 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1524.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1524.wav new file mode 100644 index 0000000000000000000000000000000000000000..47fc8b434f94ddd5e53b817136b84c634f86c57a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1524.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e52b74ecd81181f95452e042807ebb57330b6dbfd80d26670212e63a4016c0fa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1525.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1525.wav new file mode 100644 index 0000000000000000000000000000000000000000..302b7e68003a77bca22178e28fc80d5ba083935e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1525.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ff063c75af273d1f5231f133983347114a80e2e00fe77835c5c5fb2f0fca95a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1527.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1527.wav new file mode 100644 index 0000000000000000000000000000000000000000..fb98d8f20df7ea36020e96a69404c49ee8135bea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1527.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bca7e2ee3ffc243a0888b9a532e8079c5177c050d33a0defad4772bad2af364d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1528.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1528.wav new file mode 100644 index 0000000000000000000000000000000000000000..704608861bfd92d697be316bfbc2a47b1d2c47f9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1528.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c7d590e5eb2bc6897bd6f22a8757101decca6d5464113a564a548a6e050f2f0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1529.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1529.wav new file mode 100644 index 0000000000000000000000000000000000000000..fb039af909bc81f00a8ec16a26fd475f59d86f38 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1529.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6cb880085db585bc43674e14e184db297f50b290c8d317136b887d7323527b5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/153.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/153.wav new file mode 100644 index 0000000000000000000000000000000000000000..b6ff6d63bf2e1b1981713849b99f9093c69feb2b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/153.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a495db64a32e70ce800cd6f85d7e8981a697e45a83f805381dd7d436242d1fd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1530.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1530.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c7047c58c4cbe060f0223b2bdea053b5a1b78c6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1530.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1e67873a6a756ad90f9d50690507ce696b6e5e93ff827c4e3146163a49ee742 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1533.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1533.wav new file mode 100644 index 0000000000000000000000000000000000000000..fac9ca60db83a373ef8b1cd53b56d3b19dd1199a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1533.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75ec3df4942e5214d414f3851b15d2909d62d3fd98e40150b78dfa5a5ee72737 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1537.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1537.wav new file mode 100644 index 0000000000000000000000000000000000000000..acb96bbbd37ac5bf7edf66629f3dfc10a3a212ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1537.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b3b2091926217eebd9e808a2be4e616f2327916e71e72dfdde9f65bc2f96f36 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1538.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1538.wav new file mode 100644 index 0000000000000000000000000000000000000000..85b16daad5f6cad7f902ebedf99ce3775b8a8363 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1538.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:475d131171e44ba84ff685759cc391bcf2984ac1156a2e33053e6b680be67873 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/154.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/154.wav new file mode 100644 index 0000000000000000000000000000000000000000..a23d7799d43244cb2879ed6f836279999f599f90 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/154.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2074c9574e9bb8e397610672116922808c8421fd6662fbd9fc0c8962e053937c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1542.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1542.wav new file mode 100644 index 0000000000000000000000000000000000000000..549182f57c31e4092b77a87e304ae4086b858012 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1542.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36f355fb02ed56642307b6ef608e6cfac49ce6f50fe3d49405368e5503a9c663 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1543.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1543.wav new file mode 100644 index 0000000000000000000000000000000000000000..050d9ffa558dd8dc2e0e6ef2493ec4a247b33aa0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1543.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ff49d4e8e4d3bebc6556fed0a25c9628a114db9ebce33b17515aa1793eccfc1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1544.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1544.wav new file mode 100644 index 0000000000000000000000000000000000000000..ac91a8fe3e1d95000effbbc17fef14441e274e82 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1544.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57a4538a34f37d4f37133195514266349914a2f16c2635fb61a23c0ff90ad0a1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1545.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1545.wav new file mode 100644 index 0000000000000000000000000000000000000000..ce2c311d131eba6ed32cfee47ed50af61bf234f0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1545.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de6b103838a4ff6452c2f7381f9e9ce2bd4418e8ec34b3a54062f5fccacf59db +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1546.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1546.wav new file mode 100644 index 0000000000000000000000000000000000000000..5d68ccbc7243b03569c3264c9ff182353de3c001 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1546.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:322190799842e58ed4c78c734c92c5a2a3e2a7e75e8f67d535510020ccdbf2a7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1547.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1547.wav new file mode 100644 index 0000000000000000000000000000000000000000..8a295ab2c28da9321adc5547d595dde2d954c983 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1547.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13e77c70b5e62a59456ab0f49988d56bac80158474640cd75d9a00eda47a0beb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/155.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/155.wav new file mode 100644 index 0000000000000000000000000000000000000000..39fae4695ba00bf295eb07f98eddc5c0c87e2473 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/155.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7236e48839d8b932fdf78f0ad5d6fc27a4d82e2f523d6d1b5a950a6c67ba6b7d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1551.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1551.wav new file mode 100644 index 0000000000000000000000000000000000000000..8cecc7d186d04e22ed7bc63ce6cf07f738321c28 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1551.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a73062c82cc93c91bd6228404be0c534643b480925c246ce5b26e85d4b681f48 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1553.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1553.wav new file mode 100644 index 0000000000000000000000000000000000000000..b178e0ed1add87e1c9a4b55741b33051612432cd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1553.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47de0fa4bb9c5dc99d9203eac7590d03cc4423465c98a29ce518938094ffcdf5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1557.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1557.wav new file mode 100644 index 0000000000000000000000000000000000000000..4cde2c2ae66f114c529d7960b22e82b2e6d81a25 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1557.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5700c8a0f88167887c7ebb203280622131716da3fb00e9d16ed3fdb2b4efa68 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1558.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1558.wav new file mode 100644 index 0000000000000000000000000000000000000000..d07aa46ea6551db70df3e7f03ee680fb47b8934e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1558.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78163c4c606fc4511ab388585740bb285e444062adac52285d423552ea1dcc90 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1561.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1561.wav new file mode 100644 index 0000000000000000000000000000000000000000..c16bdd0805dc31ea344680cfe228e1b75c884355 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1561.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3aba4391fe1a14128b3a85320a9e897e208becaa3441203c1a1da2e718cb1721 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1562.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1562.wav new file mode 100644 index 0000000000000000000000000000000000000000..308c496f520c916a0ccea770994e3e3691b252c0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1562.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d480eb1fa05fc3dbb40ba2410203bc1fa38a75f59f02ab53b42d89b89271c7fb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1563.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1563.wav new file mode 100644 index 0000000000000000000000000000000000000000..28bd7ecadc46e092fbf318c7ab6ec882609f3f98 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1563.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:648da2f85f6826b647f6da5625d941408b9acb0500f447be3bd84332fe79e62c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1565.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1565.wav new file mode 100644 index 0000000000000000000000000000000000000000..1ca64120aea94e4d25584b4e13290d9d6621ecdb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1565.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05546bcbe624785986024d9c8ebcd4b1fcc731fe2064093122ec47b38cd05542 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1568.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1568.wav new file mode 100644 index 0000000000000000000000000000000000000000..6311c3bb7f70eebc04a04a5edb3d5b553fbd4b52 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1568.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb2588c2fb63b8cd7e726f3aaad535cdc246fb0a71785c3f5cdf8746af0d37c3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1572.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1572.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0c682d8a3646c02445f766ebc5a2ece8e88e104 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1572.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8812d3df23a0d3f6535e2f7bc0914ff29fdaae43a1923cb8b32f139308f1e909 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1574.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1574.wav new file mode 100644 index 0000000000000000000000000000000000000000..b87d99bb03d185cfa13eb4571486404712e6d6e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1574.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:427ac0ee5f644d65f6af92cc966064e8afe732e355a436080fbda8db6c11b9dc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1576.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1576.wav new file mode 100644 index 0000000000000000000000000000000000000000..47d7200a6cc4f5589d8f1aa2a7b49e0aedfe3a90 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1576.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:261bc62c6dcbb3b429249357497ff951a1e35999317f9dac45a9468b72477d30 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1577.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1577.wav new file mode 100644 index 0000000000000000000000000000000000000000..58bbc3df198299a5b3a6055231f1a07823fcf734 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1577.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25d7909c1347e30202af3c2f9ea4a161584937f95082dfcc1b84b3e23c1fd8a0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1578.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1578.wav new file mode 100644 index 0000000000000000000000000000000000000000..86eb16a8c560db017dcc8639bea73e3f5e04d7ed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1578.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ae58b475cf620ecce931762739d2549b7816268ce16bfd18323102c59e5702a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1579.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1579.wav new file mode 100644 index 0000000000000000000000000000000000000000..870f81a0a81be18e28adf7358b8a10f4d20fe524 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1579.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fc984f4a41b0cee1e554c546bd5a0419d1310d0dace8e371d76650c507dfc78 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1581.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1581.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0b0d28307b8f886ddb8377435fedc30c01503f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1581.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8a8e2fa5dd4a6c306bd91efe0bfaad84faa80f5d47297022da12c1e9b0ed674 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1582.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1582.wav new file mode 100644 index 0000000000000000000000000000000000000000..63be14526b68b03590dc62dbe8fa7112474e0277 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1582.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43487d719bc85497b9139ca9cc629f6479473c670cd48714d82756269c86d1c3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1584.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1584.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc616595548445b7aa62d9db314a82b97479c22f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1584.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0094da3ad2c121fa80b5cfa654c9674a00a2a8123cddeff3d5bf2f36ec2873f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1585.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1585.wav new file mode 100644 index 0000000000000000000000000000000000000000..38f57f1df6a6db6cfc5c85f2d8a677d56c63ba4f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1585.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d76077e182d8ab02ffcb9b703353808634c3f4e3d7a1932655f513cc2af9aad7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1587.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1587.wav new file mode 100644 index 0000000000000000000000000000000000000000..df0441710f82af8463f72a51ffb7304355176055 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1587.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4dcd1ba0a6087ab2709cc0b311361cd69768a6604011ce9cca4a1b8759745e7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1588.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1588.wav new file mode 100644 index 0000000000000000000000000000000000000000..d72d560df239ce14fe4de6c5f04210dc480d0a73 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1588.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acf2dda0c1802af8402ef0cfa2d4299a4ac47f8b5523f0c7a6e328bbb22ef9f9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1589.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1589.wav new file mode 100644 index 0000000000000000000000000000000000000000..17dcc0df6651b3467577042a4fc0c9e8724021a1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1589.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5684fe579baadb679e7b933a1b0619961bcc1e63c5be49c2aeadb14cc3e69577 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/159.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/159.wav new file mode 100644 index 0000000000000000000000000000000000000000..0f3b4dff04252a42ad3ead07f0a0a8f76ee006b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/159.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3df40f379871a825c88b833b134e2dd7dc4f367ce56f751fc2e61b7df342411 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1590.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1590.wav new file mode 100644 index 0000000000000000000000000000000000000000..295d5404f46713dc48ec36253c476949d578ee7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1590.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e6b3bd110a134869974a428303ebc95c042e76dcf24b9fb47ceec03d0196bee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1591.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1591.wav new file mode 100644 index 0000000000000000000000000000000000000000..4432924d7aec9ac199be0ca659d8521a0e72b4d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1591.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bdf6298e3280e2baedf29116b939b086ac840cf17a8e639179695b078f6e42b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1593.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1593.wav new file mode 100644 index 0000000000000000000000000000000000000000..1e3aab2714eb1a5b2285aa651b6e4f00828363bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1593.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0ede1ade2a10d03694fec067e8a719df12a38af15feb2befe8f7dcc81d54444 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1594.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1594.wav new file mode 100644 index 0000000000000000000000000000000000000000..3214abe5a32ccf30ae56bac0b5f93471d23ebcf3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1594.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5df91e5297ca3f90663e8d809873342213deb71fd42c9b624085c284fbcf9288 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1596.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1596.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a1ad2cd45635728e1695a5e2af4cc441ffd3d30 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1596.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9b457bdd38b5f162cc5bb72bb8cf58552904dca7d9b9c79c29a084c47c0c921 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1597.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1597.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe0b6b08b28e3f3951ebb04910f6cf5ac6ff0272 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1597.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff4cd676f7f8ffcf7618682ae310118cc53190e528b614a12b2e36888bcbdd2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1598.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1598.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c0caf6ac42272f78ce041553904f4221fc7e5ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1598.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:827ec40103bd5bdbb99c7907fcbc0b2daa134982167730ed264e8fd4a8235f62 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1599.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1599.wav new file mode 100644 index 0000000000000000000000000000000000000000..33aab9f40ad1f69e604344378f869d90b7baf3cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1599.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7857dc843a054b6c57336b5c820298df33629cc5e3cd809c7f31cd255057c23c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/16.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/16.wav new file mode 100644 index 0000000000000000000000000000000000000000..728781bcbe3f75ca11174230ceb0974e87c205fe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/16.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:364f516294dc90c663c75e94e3c5fefdb745c7d628142a6e68e2d8a7437901db +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1601.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1601.wav new file mode 100644 index 0000000000000000000000000000000000000000..59204b09e0dba519b0226d2a19457e0bf6186b33 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1601.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85079adc474df196bdc6bfcb2c3d01fb778769b67203117268ba95b621919928 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1603.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1603.wav new file mode 100644 index 0000000000000000000000000000000000000000..216fe82ffa95e40b5275967bed62efe70763fb55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1603.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9c832bb7b3034ab83338c8d85eeb2bac0a2f4cf5928b1cda9e21ca9b72da6ce +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1605.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1605.wav new file mode 100644 index 0000000000000000000000000000000000000000..b2373f942eb2bdf56edde5de4c70720ada5d72c8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1605.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b21e49c5fd2d3ec176c23a99fff94d6764ddecd584642dc1384de20d0cc0907 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1606.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1606.wav new file mode 100644 index 0000000000000000000000000000000000000000..e4feb2c2ce588bfdc81e351a2e072702f877cef2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1606.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cead44a3ab179137d5c00e6c3f89b1a2190490f28a240ec0f36a9ab5a88ce88 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1609.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1609.wav new file mode 100644 index 0000000000000000000000000000000000000000..c537f9727da1a3d41c96fcf130f71ea671d36798 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1609.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cc6596028803b2f004f6674d11344e31c29d511592908aab79feba668caff7f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/161.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/161.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a78d4b965030bfba2c5fa5eef2e454d9a511ed2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/161.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeaf78e1a7c4629e36fe5d0ff91a8520ce0dd417953063923e8110c5177eaae5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1611.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1611.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ec1de34ed3ed39e33687b48c827d082e1c216ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1611.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a18a400657405d3a1bc19c90b764bc33ceea3e3da53aaf93767e99a235285ff +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1612.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1612.wav new file mode 100644 index 0000000000000000000000000000000000000000..2a9b6bf5a24f5c3509cbdc2b559faed07916b338 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1612.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d06fe7f9f25d18b7efb23355eea981420623ea67c3078da9f286f29995637d5b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1614.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1614.wav new file mode 100644 index 0000000000000000000000000000000000000000..b54f0e9293bc398fdea633355663440123ac4bf8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1614.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c7a434a41e3043e615abdd8d64bc2b16053424aed9b8e4a02b2b6093a8ff766 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1616.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1616.wav new file mode 100644 index 0000000000000000000000000000000000000000..eae97c2d7e31129288ada533d9deca0d8111f8d9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1616.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b41ddfbe20866cdc3b062ff5be44cb2acbadd6bec01d29661c0f78060f8ab985 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1617.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1617.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d31f89084e72818476f50d0559b590fb991acca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1617.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01a3e338ac9bf47c85d4d1925c9aff19ae03ff98aeed6ee66f05f22ec0e1690e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1618.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1618.wav new file mode 100644 index 0000000000000000000000000000000000000000..4182db44e84c7bb14cefee8367f22c57f0af60dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1618.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61b01f53a68ee4dad57addd36fe4a4bdd209297ba90d7472bec8f82723ca58db +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/1619.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/1619.wav new file mode 100644 index 0000000000000000000000000000000000000000..4dc8bce68bfbcab7ec1acd380d45027deeb71dc1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/1619.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c5b6997f8555a1359a22707f9daaa2ab3d5b206d9d0ff9a85793b29d7f0b079 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/163.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/163.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc062b7686e60311ee3e5233da057f62b622f5b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/163.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:500689fdbc92d4f672f306d18801a9495d022be6a52c8acc74b738e9c76cac2a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/165.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/165.wav new file mode 100644 index 0000000000000000000000000000000000000000..e1f55da8e5074f6572a96e306c848657a79b2de4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/165.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01bc3b0f2d6e7c3a4ada8bf4e1e519eab25749ea062aeb920e6b15ae9fcddfdf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/169.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/169.wav new file mode 100644 index 0000000000000000000000000000000000000000..3dee5867b2a9df7072dcd1ac109efe83575d2af7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/169.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91f4cf86d3c4a46bf1340e26b414ed1cbb6aa21971154dfef256e762c35624eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/17.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/17.wav new file mode 100644 index 0000000000000000000000000000000000000000..68c05d4c48062df555a79b740b651ffe19667751 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/17.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeabc16ae1a05ad04c69eb1d7a9a91847f3c44e2d6f339a1af23ab038c8e96fe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/171.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/171.wav new file mode 100644 index 0000000000000000000000000000000000000000..d925d466cca7d4172c7c08bbd589832aa7905a7d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/171.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2c2371b3db3c4d8d13ca7b89e06bf6409d385688a8f294066af6bf11490b272 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/173.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/173.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ce1e7ce8d498b615c258871959ac11d4c6173b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/173.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc5a2d893c0c6be7bb2750024a0c9b225e3fd95e274500dabbb7bee0f0785ea6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/177.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/177.wav new file mode 100644 index 0000000000000000000000000000000000000000..d841dca122698c5c4dd717f2c147948d8d80255e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/177.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84446f7adbef2ef5f9011bfb38d170357351d8138eb5e92c06c4b1e87ccdc002 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/179.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/179.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fdf7a7879744e670e3783846fd5c5312688e7f0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/179.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:150f48fe0f35e722438dede513ba67664c14fa71153c8e20b909e27b3109a9de +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/18.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/18.wav new file mode 100644 index 0000000000000000000000000000000000000000..7807514c5d2184f73acd23fa590e28e62c95db67 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/18.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2a1fd14ade15a7323ebd7fa19bf15d2815e0b5601df03f3b0f2ff9014abc0c6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/182.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/182.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd4c80dc60ff318b17a170e9428dcd4a5ecb80d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/182.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a140e8a082e4f98441a00efe75984f87e257b3193aee01660525952c1d8ad10 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/183.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/183.wav new file mode 100644 index 0000000000000000000000000000000000000000..290f14150d8dc865cc5d7afe0372dbf504f457a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/183.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fcd9ec435d3e3866d63d8e59886bde20eac66359e70b99a3c33a71506ec5d99 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/185.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/185.wav new file mode 100644 index 0000000000000000000000000000000000000000..1bcabaa4d25f7f263aa3bb20027f3b141d35c103 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/185.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22e75dcd45a864e6c3974bc93c46eb56b63733136624228620a59febc01636f4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/187.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/187.wav new file mode 100644 index 0000000000000000000000000000000000000000..6f7213c17eadd1ab34462e0f899b389907748f3e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/187.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc9034bed1733102bee3cf35e9485e342b7d9c19f928ca3baaa28645984c8134 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/188.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/188.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca6c7e9d431543b6ca3807de1bfb8d77bd02f3ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/188.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70ad3554bcbe78dc02c0744551b04180bcc12ad2b1030060c7a4408b27443413 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/189.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/189.wav new file mode 100644 index 0000000000000000000000000000000000000000..7837e4e88e9f129c67c709444b283a5ed5a5ac50 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/189.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd66d5a0ee956ba8c13992f778a9face881d3187120be339862075f4e9c77c88 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/19.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/19.wav new file mode 100644 index 0000000000000000000000000000000000000000..087d89454edb3d7b6b6731e93b4977858e07e657 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/19.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01dbb11ed9f33fcb8f7adad43753870bdeba45dad2cfe66107ce005dedcca4ee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/192.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/192.wav new file mode 100644 index 0000000000000000000000000000000000000000..7eff1589e73fa5a413003c3f996845c0f1894fe8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/192.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72d55c22195f2eedc5cefbb2345130081419a70b8b392b15902be4c3740728b3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/193.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/193.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc1b70f9da73bd0f71814bd5003a175f126089cf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/193.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b58ec1c64cf309d8ee48afb6fa91d45df44da0fe5999c407c0c5798ac869ab2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/195.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/195.wav new file mode 100644 index 0000000000000000000000000000000000000000..b573a0d62a23b9290d4985c78ee15b6af1c2267b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/195.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d44510a164d841277d2b1f4d9ec67cb91ad8912be50c00b0ec33d867b7c71759 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/199.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/199.wav new file mode 100644 index 0000000000000000000000000000000000000000..43046fc962750e87fdc9b996bc96b2caa6e992de --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/199.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56e16f67e80a93575157619dba8861f04905e7a72bfed4fb72c8630d12765fa0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/20.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/20.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee395783dfb5c972508f31c2e054b4951a38f1ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/20.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1da7c30d38c231924915748e475a6b2a20bef1716fddfde6d962855a2e12c00 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/200.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/200.wav new file mode 100644 index 0000000000000000000000000000000000000000..45e12a28ccd7a0080463269d2ff4fae770f6fed9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/200.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc45ef4cb02379398c6af2f6d2fe17fb72d304b3faa1e4ce8b042baddaac52d5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/201.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/201.wav new file mode 100644 index 0000000000000000000000000000000000000000..b442ef72e90e05ba1392fb35bd84a19670f08f1f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/201.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8be1d2fec9bb2afa78f13f3a86de851a42124349d1b7382ffe4348a690c1b7ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/203.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/203.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe18c42dd6282e0961fa45d966ed2bcc047d3eba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/203.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5724db09ddd9bc72cb8f5c2bb78d19067828bf316a2f922b3b4e046d798b31cc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/204.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/204.wav new file mode 100644 index 0000000000000000000000000000000000000000..12f1278b06993bffe1fde77e8904ef22946e6aba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/204.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1efc96d3bde08f33ec04e6c8aa14fdfd98657e69d25eaa10fb2a6553973154cc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/206.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/206.wav new file mode 100644 index 0000000000000000000000000000000000000000..37d9da10b0fd641c02453dea272846b0fa8690a5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/206.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6eb36104575e0180def31b14adde3bd8a094bc8fb37a949c9f7ae4c03914f17 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/207.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/207.wav new file mode 100644 index 0000000000000000000000000000000000000000..f93627503aa85c4df91c8f9b821d36288e574e7f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/207.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68cafc7a74c6bc42e0a9f10d9794eb2b8b5670cb3b16886771406c4449f136ac +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/208.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/208.wav new file mode 100644 index 0000000000000000000000000000000000000000..f087f7953b3887985e7a646494f1a30168f854d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/208.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26c4687349259d875da2a6fb33c75773657c6e79e9099daa15f8ec06c0585323 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/209.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/209.wav new file mode 100644 index 0000000000000000000000000000000000000000..cad31870b03d4c895ef3f80c35779dfbf3ae7ff8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/209.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e729676395be200e2abd5db61b7a53af58d406b16e4fd74d80e2e477380c510 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/21.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/21.wav new file mode 100644 index 0000000000000000000000000000000000000000..3ecb7348145fd4524ea3afffce1c98da1a00899e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/21.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b683e98fd76b69e17360ce655259b02748b5d54cee632385665bf0b016a636aa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/210.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/210.wav new file mode 100644 index 0000000000000000000000000000000000000000..a924c1e2404433d7ba1031e3f76c3bab8b1815c7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/210.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88cecfc20f966ebc9922b26eb395b0a677190b77ebf3cbf342833a26afeb4387 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/211.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/211.wav new file mode 100644 index 0000000000000000000000000000000000000000..f669559f17710bd97ac753f24a644cab628f5193 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/211.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d94492c98242290747650d400f86f02bb8c3aaeaaa685b6e2ce6063001752a2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/214.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/214.wav new file mode 100644 index 0000000000000000000000000000000000000000..cf582930e7c14a1400126dadb8668bb7d4d84ff8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/214.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf68d8d5398420ccaf22727572abc79d48882a38ad651363e800f76d3a9e086c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/215.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/215.wav new file mode 100644 index 0000000000000000000000000000000000000000..072cab8641960417afe766c828f3b7d78a596f8a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/215.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06b495f7f3f4edb73bc5e9a5d8dfb0602ee9b900385f689b8193bd42b6e4f539 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/219.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/219.wav new file mode 100644 index 0000000000000000000000000000000000000000..3aa6d90afaf5f8bf38eb5b278e903f9f6ab33846 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/219.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c82c25e9cd463f8f5c37a76c33268ccac495d6395f50e548219fd4e26185c30 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/22.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/22.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6f43b144f2f5d5bed17f7dbd2ec0a6469efe1bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/22.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7091be6cdc756b35600055e2bef94197636f2693fe955187e1a3f8121d1f46ec +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/221.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/221.wav new file mode 100644 index 0000000000000000000000000000000000000000..3dfcd454c5bc879efda865c2cd838812b8731844 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/221.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8700303165c397d2a686578675a1b8f3223105066e948d6a0d08ebadb301af34 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/222.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/222.wav new file mode 100644 index 0000000000000000000000000000000000000000..caad537ff20662900141cbb3816636dbd365e573 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/222.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:601579e0b856b033b87a510cc3f9bac30617b005deb9fa5a8289694b4af47beb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/225.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/225.wav new file mode 100644 index 0000000000000000000000000000000000000000..982d15d8d84bf95ae38ae24ad73680547fc35123 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/225.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31badead127d9c8ff91cf364be69723ca44de5b9b83cdc8295d20d4f15871024 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/226.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/226.wav new file mode 100644 index 0000000000000000000000000000000000000000..70e133fcb015fd764c5b5d79037d70d797cf390c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/226.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b9db2d8c75823da56a5a4901c0b88aac9a217f51f0a87bb172b4ad0b6ef04a9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/228.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/228.wav new file mode 100644 index 0000000000000000000000000000000000000000..7e1fb9c4fd50c631b7b955cacba32e04eeaa464b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/228.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecbb9448163c6b5c83c40a2bb4ef591c0e161d4323b8fbba3daf38c60e95dd40 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/229.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/229.wav new file mode 100644 index 0000000000000000000000000000000000000000..0393d634a836e066c4df77f395e01bac1d791df6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/229.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b59a038f928a1aa189d1a2ea5ee337d743f7776d1adf5ca6e2bf4386785548a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/23.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/23.wav new file mode 100644 index 0000000000000000000000000000000000000000..354cfa422a7514138db3d9dd49dc4e3b02c6b184 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/23.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04c6b433f04e3f552c82bed61d1954d2a37f3031d9f6a20b619a8ae0da6f1521 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/231.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/231.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe0ed6d238fcf714393e200ee3d0615dd4bc76c7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/231.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2b4df12ec1ffc1f351c96135c7ad35b4c45e27589a329425c3a1782634304c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/234.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/234.wav new file mode 100644 index 0000000000000000000000000000000000000000..895aaf5153bd7b7099a4780416a0ff8d90076133 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/234.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d92f6ce7cfde59906b23d98b5447828ea64d0086fb1176859b93e9a3cd5f15c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/236.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/236.wav new file mode 100644 index 0000000000000000000000000000000000000000..895026b3ab5cbc485fefeee5b0c19d6a7a147a5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/236.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89b617b1d05403a25e6636f500c90c0c19d094cfaa0f26e5a00bfa86a2595e6e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/237.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/237.wav new file mode 100644 index 0000000000000000000000000000000000000000..7d34b465d7e4e782a055db6b0ff3aade799f284b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/237.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4b4556912d7bc218900baf1876eeec3c66f2d96dff1343aae237abb9658a5b6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/238.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/238.wav new file mode 100644 index 0000000000000000000000000000000000000000..7e0aecc595147e7c426841deb5649e53322366e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/238.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6af8b8b52101e4289be787a2e5a48fe9295a42fee6e16178c82c926b618edeb7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/240.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/240.wav new file mode 100644 index 0000000000000000000000000000000000000000..6839b01c478d0cbf4f29c51ef6fb5baf4c1079a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/240.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ab9118c2fd4238c830195f725b1f586bc8d59993396df605478b394aa467cc7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/241.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/241.wav new file mode 100644 index 0000000000000000000000000000000000000000..92d74e71d42981dffe7cd8f95333b0259dd2a6e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/241.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccdbb23a2dd72ee9bd833c856625dd52ccd28f4ceb1dd9ad014e182275e99436 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/243.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/243.wav new file mode 100644 index 0000000000000000000000000000000000000000..d83d63ab8f9579360c83564eb2c3ebfe55b22ac6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/243.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65f5014a214347a0e592106ff1889ffbcf2e085f1fa2d96781d0a2f66599fb29 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/244.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/244.wav new file mode 100644 index 0000000000000000000000000000000000000000..93400d7a731c52a5099ae42f7626eae5314dbd7c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/244.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99e12e3a3e4de7b7fe9c9b54de1c41a93aa8a8bd8ecffbed1d6cc7bebf7b7431 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/245.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/245.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ad5634acd23b4451858118373b8238e436e528d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/245.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c70e8a68b512ec9e727768f6ea48a7ce785375277f58dc8e21b25aa29b8c18bb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/246.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/246.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b45271e6060674762553218d2212c72efd89d51 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/246.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6441380c6db90c4a7efc55a3c0b66a8c5d29a098eb08f435d58a4fe21960b52e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/247.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/247.wav new file mode 100644 index 0000000000000000000000000000000000000000..1259cb8ae8625c485c1b93b7b2f6819972671b6d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/247.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b28a744cc963acbb902edbba06c946765b20378cecedf5cf257a70a6b27281e9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/248.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/248.wav new file mode 100644 index 0000000000000000000000000000000000000000..828f3314f87518fc8892d2434e2a075d2e40667b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/248.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:607e35d8ed73d963831ea1e03a8d269281a2d02a6f950923a733eab69149973d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/249.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/249.wav new file mode 100644 index 0000000000000000000000000000000000000000..31a5488894ddb3639243410daa82f0d9e7c99ce4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/249.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7ebaa24586f7917b2c45ec3a5d3f109fee7973f4f4a12bed0b16e1470c6759d6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/250.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/250.wav new file mode 100644 index 0000000000000000000000000000000000000000..95601b7de2933a45e67205270970a38d3d5f4994 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/250.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fd56118a8daf180e4de07f7ed725ce109b4ff7f207e665e1605cfbc041d6934 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/251.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/251.wav new file mode 100644 index 0000000000000000000000000000000000000000..0c1e7078c35f7cefd1ffcd8db8acfe5cf1461919 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/251.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66fa2dfa48fd62c5cc87275381dfcc2b39c191200d93c08d9fc2cf50d66abbb8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/253.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/253.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef1b8bcc9728addeac5ca90fa8be0913237216b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/253.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f61eb13a738a44777306e16686a5c24ee90f2691e8725ed7387815afd75a446 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/255.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/255.wav new file mode 100644 index 0000000000000000000000000000000000000000..632ef4f8611ae83aefab1c9cb22b794a6f2c0372 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/255.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:014ddbeccdd77221304a372728cb637e070420bf93869e25221de9b197912610 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/257.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/257.wav new file mode 100644 index 0000000000000000000000000000000000000000..658d4c4c69168da79193692671bcbd16568ab178 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/257.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8885c82947d2e092ccd4f8d4e812b7e2391313171573b8b73af8ef6ef000e6ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/258.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/258.wav new file mode 100644 index 0000000000000000000000000000000000000000..43d422d9faab4a7af8a3873660fed381c67ed4cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/258.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e14b9c7beb9a525c841053cc116f7428b28070ace55b0bdd0d23f5643d413b79 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/259.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/259.wav new file mode 100644 index 0000000000000000000000000000000000000000..043d16907c205af3b735f763dec77d4774a8bd19 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/259.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a60c31dbc9f19dd7912f6a07d71ebef69db44e5996700f6e28057c238bd1a777 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/26.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/26.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf65e2369d31060babb359dbab4f515ea130fd58 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/26.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e9f0438094d10c18d04c493e3fb7eafca2a0c7c008a822bbc7adf459711ed0a8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/260.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/260.wav new file mode 100644 index 0000000000000000000000000000000000000000..f1e026b31f77959cc724f9ae6515316f207e46e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/260.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fdc1edf0b3d9259f5f849532ad90acfd26c13ae5e7fdcace76a9f1b2feeaa828 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/261.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/261.wav new file mode 100644 index 0000000000000000000000000000000000000000..68ef1ee36dd6467e241ced07e49f57c71c455c71 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/261.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ec0b337dc21a71ea26633ab01dbb04d6f39f2d1cecb8555e9dd716f8dd6df24 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/262.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/262.wav new file mode 100644 index 0000000000000000000000000000000000000000..454876550604a153b624b1aec7c57f5af8b4989f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/262.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88b54f2b3d6754db4b7b8364619fa68047990130190e7020c0c7b134ab189730 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/264.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/264.wav new file mode 100644 index 0000000000000000000000000000000000000000..464410e31341a712ec4e5f6a054202e555507348 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/264.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e5af1f04fbb3fec319ad5f3d2c7705fa8e0061067016ebef1c43c2dc6892cd1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/268.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/268.wav new file mode 100644 index 0000000000000000000000000000000000000000..6fc366214233b6ca342b490c9c7fc68cda9824d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/268.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f7c2bac81c94f2f49933debf7cae8946070d596c6477ca1bb1d06b90fa8b284 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/27.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/27.wav new file mode 100644 index 0000000000000000000000000000000000000000..10d1fd7dff9d0ad0a54c1d4adab33a522d7c9adf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/27.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c04c0ba29c176e67fadb81e55802fa5c748eb76f44a4eb68948a4e25f3330c4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/270.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/270.wav new file mode 100644 index 0000000000000000000000000000000000000000..bfb181dc8160655b0ba0496c75c8151e0ab87b84 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/270.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1771726c351f74e2828b000748309819a02c044415f1346b4747d7f7faebf49 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/271.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/271.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0290b3634045095393e5071047c2eb9212e866c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/271.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5078288a7648497d7ae46c2551bcbb62c8f58a7b9648e8dd00826f88bb8b0557 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/272.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/272.wav new file mode 100644 index 0000000000000000000000000000000000000000..c48c6b880da0671b50e4a4f00f9d2531f0153e14 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/272.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82458aa3525d1e4ed4d063da7028664e81d045eb748b12607d5d963e28b4c0a2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/273.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/273.wav new file mode 100644 index 0000000000000000000000000000000000000000..b18aa31d77b21918c44451ffab47afd788d5e30f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/273.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:128263a47b7d577df8a7cc8b1cf88dccb7b64cf3e9e6de892c07ee3f4105713b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/274.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/274.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f1f5f6e0f5a9872aa3ac47feb40c136b366ef9b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/274.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b70dbb3880ed608c2fa24742dea845c1ce2907e7ba1da7c88e19a331bb3482c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/276.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/276.wav new file mode 100644 index 0000000000000000000000000000000000000000..fb74d18ddeb380c1c2bd4477b57c584687f54f9c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/276.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7a626fbbc930d0d9c0a8a381b9cceda7ee41c16fe1d2c5e38475fc9d8bcef8f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/277.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/277.wav new file mode 100644 index 0000000000000000000000000000000000000000..1614873e3437a603b75d0db686e81f8d2bd27b63 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/277.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0c6e25b27319650bdd077da6e4224df1d89075497f2fd3264f1d2d0ed4b3f2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/279.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/279.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdb403ad807baeb931c673119abe6c359756f914 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/279.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1651b9b8be32fd61ccc2141f8526bbb1d5b58e58f4e2554582bd7307f486aa1e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/28.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/28.wav new file mode 100644 index 0000000000000000000000000000000000000000..f59c2de28f5a3399129140c0c43550e0adfd0523 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/28.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fab73a23b1ec4dcf7985c5099b11aedda78f14be31d9b453d169bc1302dba2cd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/280.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/280.wav new file mode 100644 index 0000000000000000000000000000000000000000..c39627ad7057de7cc6ead89223d5e380a37f0b05 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/280.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55d752496887c1381dc419383ad4aadaa9474de43f1d604fe0920933808bc619 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/282.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/282.wav new file mode 100644 index 0000000000000000000000000000000000000000..bb8be7fcb7d406255df244e29db0fa658a19a165 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/282.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe1be05c4ac169152f5a85c6c61b14bac101d02317b60bec768fdd2ec12bbc99 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/283.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/283.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bd571ee46c2403ced40975ec76d5b52be9cf97b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/283.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:561c4ffdd2553784566058d4a36b57c5ff0d6668d5c3b8b7b31e328e2ac83a2b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/285.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/285.wav new file mode 100644 index 0000000000000000000000000000000000000000..4a42a0d5bce9ea2dcfb0b67f6e83992aca3425ac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/285.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20f6587bdd76ce2db45ed223781f0227e6946fb7f497b533959090301209d99a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/287.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/287.wav new file mode 100644 index 0000000000000000000000000000000000000000..95674eb36f2433ab0b3a78309e9bf996d4257b2e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/287.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23d15e9190c64b0a761bd42bb198ec787a10792a007dcbf7de9d64c132631cf0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/288.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/288.wav new file mode 100644 index 0000000000000000000000000000000000000000..47472894116a07d84913781f86fa17c727f99faf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/288.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a988cb7d6bb68a756bc9be9b53761b58c0272503696d1186fc77735d8e4e1a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/289.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/289.wav new file mode 100644 index 0000000000000000000000000000000000000000..6790f5655cd3dde4700ef4f5361013a8aa968f67 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/289.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35f3d2b8afad1dd15e05f10d028399be22072c0f027b525d89b839d5d761c0da +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/29.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/29.wav new file mode 100644 index 0000000000000000000000000000000000000000..ffd49d1ee553d98c5452761d24de80f1a9e179a4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/29.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d48031dfb999000a54a87bacbe548c2282a07ddbae13a598f5895310afdd523 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/290.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/290.wav new file mode 100644 index 0000000000000000000000000000000000000000..4f81363cfe2eba73328465b5e6c7aa8d7a02cc39 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/290.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e57ba97b0f27ecac864807fcc5a5f2382ff492178612dccc2cc632e906bdca70 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/291.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/291.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ef70f8a4c6d619ca3df20b09a247bb8a024494e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/291.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45b7219a284795ac90a980894918b34773d10f22f784d9678074d7faca09e48b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/292.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/292.wav new file mode 100644 index 0000000000000000000000000000000000000000..8de5ce0e4c042d33e279819f8c01139eb7d19d77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/292.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:433eb44dd05707608ab44f6fef4354b5ab6ea42bbe7357b5cd5c092ed2cee01c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/293.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/293.wav new file mode 100644 index 0000000000000000000000000000000000000000..53a35022323ac58ba59f1c7dfe117376cc7cc595 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/293.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b1e179720028fa7d24469f2419c752df2b695625f171acb323498415844e058 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/295.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/295.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1589f908de995013f1f9e56843eadf6a44c06ce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/295.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab75217453cb695cf8e299258f391b1033f9a1c4eeea7688acac7be00809e4ad +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/297.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/297.wav new file mode 100644 index 0000000000000000000000000000000000000000..78114fa8d826fa9bacb4597ab355500828e6b9f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/297.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:309d0027125f8e01202c619b274cd23b8e7f5e840081f0f78ddb5f367ba5b89c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/298.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/298.wav new file mode 100644 index 0000000000000000000000000000000000000000..595fe2e0b7acd10e710a801efce236eb25d571ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/298.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d532045e815c66e9cde80ecb9bb40e5efaf45df54077d60d71ea532ef66d921 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/299.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/299.wav new file mode 100644 index 0000000000000000000000000000000000000000..2bce20bcb59ebcabd6d8e146a13d2b2b75dca318 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/299.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15fd81fc2abdc521040a4c96d0dd5c74ed90b101bbacf05a8c08983b8a126243 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/3.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/3.wav new file mode 100644 index 0000000000000000000000000000000000000000..d98fdd6cea5adfe9ed750d5064842a700c9ca2ec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/3.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f4625e1590fe07c2dc50574fb8f28c0981f1dd0750035ad414e4000c8bf05b4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/301.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/301.wav new file mode 100644 index 0000000000000000000000000000000000000000..5e621bbff9f49ef6db0e181dece75fc454203e65 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/301.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e4999d38a3fc7dcdb7c23a8d347fdae8f1aacaa44fc54ebf5dca1253ddcb91a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/303.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/303.wav new file mode 100644 index 0000000000000000000000000000000000000000..dc4bbfeee71fa7a19d27d36c15ecaf52233e90d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/303.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:380618cf4002ac63708be19636203778c6e7058f7537d5c607ec92c99f972ac8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/305.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/305.wav new file mode 100644 index 0000000000000000000000000000000000000000..a4d5621479fc66f3df845240fd2ae0463abd6b52 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/305.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92195d898c57ea2826128f9a720de16312bcd12cf26d4ea078d2b0902c8aa101 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/309.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/309.wav new file mode 100644 index 0000000000000000000000000000000000000000..c35165e3f72c18dfa484906bccb41b6bd27b57a6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/309.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c544a42d5335d2cddec2de67f2976a226dbddf1b90917fbfc6dc713e66758db +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/31.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/31.wav new file mode 100644 index 0000000000000000000000000000000000000000..99b66c23aad1f285c2b3dc54d16737104bfab360 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/31.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff4e88dd7736afa7f70f43cb5a6937bc507b3bf72ebbd8ce5c52e62fa11ccb7b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/311.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/311.wav new file mode 100644 index 0000000000000000000000000000000000000000..15835b4a008f5141b5daf05abe3e52ac14d1f636 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/311.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cdea33892605d9b0b0403db39a22bba8daa323c62be333d125c233f2ebdeb31 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/312.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/312.wav new file mode 100644 index 0000000000000000000000000000000000000000..2adfc0d5fa612b296634788ed1940975a3e0e313 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/312.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc72522fc6603ff7d43b7520ed74ae776bf47b44103c66805557fae2460fb735 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/315.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/315.wav new file mode 100644 index 0000000000000000000000000000000000000000..31bda35ab8c0ebf871bfc4a984ec2ec348120bed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/315.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dab845fe59b05c24b4d899106dee771d03b508c3f462427ac8a4f34871090e5a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/321.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/321.wav new file mode 100644 index 0000000000000000000000000000000000000000..cf17ff9d5aaab13a459ecf00f145c47ceb621818 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/321.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90a4af63552949284b12ce63571257997fdac2b1558b0cca144e969e00b087b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/322.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/322.wav new file mode 100644 index 0000000000000000000000000000000000000000..31d136c0f78b0e6c7200ec76a1221df046226b11 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/322.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ca1bae6551ed5d5a7b93d3c40475a879a5820616e630a7d29f3be4ff8784e1f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/323.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/323.wav new file mode 100644 index 0000000000000000000000000000000000000000..570d235ffaedf456aa68b1fa7c8a93825f7a33c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/323.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc0efb3df1e53ceacf353ced825bcf98f012d1ea655087143620fca94afd0d83 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/325.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/325.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0dd367274e8c917adfba0e734e0a83d041435af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/325.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f99c58bb56d8b92c9a1006603d93536d23d77dbed4330e80e0ebacbda7adfee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/326.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/326.wav new file mode 100644 index 0000000000000000000000000000000000000000..f08f4e22fc2413267fd40500ff5ad065247d2c89 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/326.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2dab6c7b05618003e13125197f938cab68b23f027000fea4099b09e9fbb9949 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/328.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/328.wav new file mode 100644 index 0000000000000000000000000000000000000000..033791d7d80578ac7735e848774298c63ac2d277 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/328.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:265d3d82526383c13c3d40ff611f4c0ac36b2c09e0dea8abc1291b60b3387648 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/329.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/329.wav new file mode 100644 index 0000000000000000000000000000000000000000..56f52df747a11cd8b46a5ed4604f6271b48b89f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/329.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a44e110f72ad3199a7ea208d1b2fa6f9c9d2542577acb3337dda17ca98786364 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/33.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/33.wav new file mode 100644 index 0000000000000000000000000000000000000000..9bb63f7f396b8091d01cce715a143c823ca6dd37 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/33.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeefee05fb139192e2f4cecdc707e798c69957eb4e7fa4374bf8d31aeb0814b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/330.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/330.wav new file mode 100644 index 0000000000000000000000000000000000000000..62011188b1366b6d9016cef69333d3aebab8e722 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/330.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76340ba8054ae76968868aa07c333e0ad396b5f85488192421da375629293dfa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/331.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/331.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb8648a7f12933b8ffa9b0ee01e4c894d493644f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/331.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de14d42c7ca698c8a60cc24564dbda16cd7396ec461013a413f335db9e2b402b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/335.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/335.wav new file mode 100644 index 0000000000000000000000000000000000000000..84e2ade432e8056ae4c22a145446d0e1b5ecd9ba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/335.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c5031078e40a9fda00d900a51c36821d48f93b6f358cad661c539e1830a8d62 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/337.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/337.wav new file mode 100644 index 0000000000000000000000000000000000000000..d9c7db5ebca7a96c04ae33da74a44d0d5d1463c3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/337.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b9bd46d256a93e9f81e54d0ef5f8dc98b85e1eff7dc8a5cd55eeb560dcc537a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/338.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/338.wav new file mode 100644 index 0000000000000000000000000000000000000000..d27eee21dcb0511de1e94054bad8c83b147cbdb4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/338.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17edbfa4768cee154f489f96492eaba157e112db384fec0378f2bd63fee9c0b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/339.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/339.wav new file mode 100644 index 0000000000000000000000000000000000000000..d6c17308f31c81ac4a28e5122893520f472eaa4c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/339.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bffab6e80d194b939b23224f3662cc70568acf1356eda2b647d5e1bc0ea12010 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/34.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/34.wav new file mode 100644 index 0000000000000000000000000000000000000000..b70a80c3958db4f69f683eef4d6b5fc05dddbf67 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/34.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb65785bf1ed4a00f073cc0a3e1137cd69f599c48f35b69e9a00c8bf09a3ae49 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/343.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/343.wav new file mode 100644 index 0000000000000000000000000000000000000000..26d4872467bfad6675072eb17d9b5c81709e07fa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/343.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0177e0c2ff1836f48f3b6549e6164522b0826cb3bc171d2bc0b146b896128b81 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/344.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/344.wav new file mode 100644 index 0000000000000000000000000000000000000000..8acef3773efac4f5bf75b392b3f1584be796a40b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/344.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b7860bac812402b969eeec2e12448ee167c4776c0d06fb2c1fe777c77e34ba +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/346.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/346.wav new file mode 100644 index 0000000000000000000000000000000000000000..c31c521df21dedd435504febf1d788accaccd369 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/346.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b3c017ffd0c15a2c48de68a93dc7f2ef1c7707f2cad49273de905f6fc6072ea +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/349.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/349.wav new file mode 100644 index 0000000000000000000000000000000000000000..a660eb4d1d4162237ed542f6cb2042834f68a1c9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/349.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8575627dcaead3f319d40182565aeeed451c36d6b993fcc03b5da58e6b0ae7a6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/35.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/35.wav new file mode 100644 index 0000000000000000000000000000000000000000..394478f1d4c0391aa0eab6ad9a818e91ad512899 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/35.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ccd896660327c33ccc972fb44d67dd826a8818b20c48b88b364e2752df6da2b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/350.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/350.wav new file mode 100644 index 0000000000000000000000000000000000000000..8a1446868159c5806b81fb808020f5d6d8f63b16 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/350.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:111551b0f87c984723c117e4bab0b8e064ab63a32c58eb9b986dafa68bd5b708 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/351.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/351.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ec5f8baf51a14d986d6c8dc878c0681a9164b0a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/351.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3237d3a17a59c627a1ea691904c7dd6634a0a8c730781ae581c2eca2df85aef8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/353.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/353.wav new file mode 100644 index 0000000000000000000000000000000000000000..f9898a5041066aa8382a368b1a3c673adf2ef1d5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/353.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b905d319ee07f8a72ff24d0bb09339b1e020b33e31a5fd7fa6dc426457d53976 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/354.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/354.wav new file mode 100644 index 0000000000000000000000000000000000000000..5f1f5b64663630d6eab61eac8f599081fe32b3d2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/354.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:074a82a6161345f8c8f0a0da25ba4784b56162d83484f84497c5fd04144be61a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/355.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/355.wav new file mode 100644 index 0000000000000000000000000000000000000000..b683a2d52bf196b70914cac7e379b1dea9ec3e42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/355.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5dcad2399119c2b767cf3a2a088ad504cc9c48f87d0db36b8a4c38a0f909a87 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/356.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/356.wav new file mode 100644 index 0000000000000000000000000000000000000000..a5313abc5da71d68d62d77e740a945245535d7af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/356.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffa0f2d17d5c7a6e7026d23e6c0354b3c4dff26f91aa7feeff4fd618e528b911 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/357.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/357.wav new file mode 100644 index 0000000000000000000000000000000000000000..e63b958ec298499b0abdfb5038978ab03beee8ff --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/357.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8813e9864d27c70849fadedcadfbb72e67a499f803c647ee2563a3d65911985c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/358.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/358.wav new file mode 100644 index 0000000000000000000000000000000000000000..4e90a81604df85b5e0ba64e8c73ef3f365bde1c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/358.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1143f3fbc5ab6818365ebad6163f5a904baa0b4fd344b94ccf3b8b66cf9a2172 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/359.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/359.wav new file mode 100644 index 0000000000000000000000000000000000000000..31d92afd8415158ccac3a23aa9ac4f1cfaf5a672 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/359.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32e36e7cc6b885dd06f5d81e7e519a6aa2395416eb01c34c3c045b0287190afd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/361.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/361.wav new file mode 100644 index 0000000000000000000000000000000000000000..557266ed35a31b6f7937620d3b3289d97cab6622 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/361.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9227784c7bca72cecd252304529433bd105a52e197b2fad7696633c919bcfe8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/362.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/362.wav new file mode 100644 index 0000000000000000000000000000000000000000..86509b18ce25e4bbb6e5b282ea691a8c53fa3341 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/362.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7aec8ea50edad1a313db7ba254d34f7b32cbe4aefd014497ed58a3288322880 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/363.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/363.wav new file mode 100644 index 0000000000000000000000000000000000000000..74bcfc1cbbcae4437b340249412751af80b7add1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/363.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34d24bf868ab093eb59bf217a8bf63108bfad2316474f1f67f9460707908b1e3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/366.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/366.wav new file mode 100644 index 0000000000000000000000000000000000000000..9feab685bed689fb2408e9f6d1cab6cba836dec4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/366.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:995a85280311252dfb75bacf614f4a39f976694e069dd369b6e0e3fce23ad614 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/367.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/367.wav new file mode 100644 index 0000000000000000000000000000000000000000..363254f430220c22cc30ec73e6ae4984b13f27b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/367.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dea63bcbcdb06d9fbdc24281da38c19288eee2b23419948df2e5a12be8cc30b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/368.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/368.wav new file mode 100644 index 0000000000000000000000000000000000000000..dfd254dbbcf15c4e5cfe35990a468106c1a5e8d6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/368.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69b6782e432f653f1f1c4a0f94737bd89e25d48d89f020edb2e346dd236ad867 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/370.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/370.wav new file mode 100644 index 0000000000000000000000000000000000000000..c8342d24aa6436a028d16e4edb742ca2573c06a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/370.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49c9244c5075dd1cbc7fbb28d6d3892a55b2774437dcad69df274997a5d20f8a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/372.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/372.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea9b1184c99345884735c3eea0263d5140b386ba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/372.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea989073d9edf17dcf4fe0ebb5b4f651d609c6f627f5ce3811875ebe4cd94182 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/374.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/374.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2fb95019d77b9d0a9ff3240735605f7c2b759a7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/374.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3d8af5749ebe56f963b485c5f22b25754159a8357473dec6db8d2418c773399 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/376.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/376.wav new file mode 100644 index 0000000000000000000000000000000000000000..fb6e1a05ee91a69330575f1ce5d6f29651d01c0d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/376.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d06e816016e8977ab32d0eacf2147cf2bb39531a2c8317601c910a44b96d26f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/377.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/377.wav new file mode 100644 index 0000000000000000000000000000000000000000..5005c81e4a2d714a8ac51b50ebb539d29b72ceca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/377.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:256c4a499daf64c92f695c98e5eba9ad97d68d0e52326eb3b17a0531668e6fcb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/378.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/378.wav new file mode 100644 index 0000000000000000000000000000000000000000..504a6aa7ed3eef6fb1d992eb03c73ee6a209e1f3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/378.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d5118d075f16d042ee0cdcdf26b0cb0f87b09ff300cd8a4fc4bad035975964d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/382.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/382.wav new file mode 100644 index 0000000000000000000000000000000000000000..521c3baa95a278dd5207550ef2d8c4535a23e7ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/382.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e885ec80507ae966f82e04de9a2f9e868b0560420339c86b79abd00e05d2da56 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/384.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/384.wav new file mode 100644 index 0000000000000000000000000000000000000000..8aa69458060e978c542b6ed70a072f8c4d412e5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/384.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d75456015e5e858425cfe1d47d6ecf82f93dafb684083e8bab53f3a0239114ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/385.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/385.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b639e4130b6a191025c2a90f75a760f6fa66e1a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/385.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e498c3d93bf9adcd24fc84b27d53d5b889183d85d3e1e625dbaf11f7c356e82c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/386.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/386.wav new file mode 100644 index 0000000000000000000000000000000000000000..520ed7df08ca7812409de9cc00877ac776faf50e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/386.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c88130d167e80026759047e2d671f5b9d583ee828b7e9fbd023ce67efb60eb8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/387.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/387.wav new file mode 100644 index 0000000000000000000000000000000000000000..9fff238e0b3493aef3c7294ed7f2f2e31ddce7ce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/387.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72cdeebf5daaca82e2207167edf17d6396758c2aeb55d360b0d803a607050643 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/388.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/388.wav new file mode 100644 index 0000000000000000000000000000000000000000..d580b84e3b480a0e0e12986a47b4cabb9e084554 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/388.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d34d87cb5618ba3b4941d02680e7beb2307949fa7dfebdf5492b2e46691bd4fa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/39.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/39.wav new file mode 100644 index 0000000000000000000000000000000000000000..2c32526712dbf744a2e409f472f99618bb9e6565 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/39.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47b069883fa0f3efe288db8c30684532692490ce69beb50391994685b2273fd4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/390.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/390.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef6ef6a0e41858df6b04b410ea6c29757c388f92 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/390.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d558599108b97048e5ad30e13fb84b225df72ca4f8acd02a6344b60a59ac576 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/392.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/392.wav new file mode 100644 index 0000000000000000000000000000000000000000..9407d347c460f0c7da992230506b3a6c0a37fc24 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/392.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3e42e3df5ba3ef414d1bc48dd44dad8dd2ad13fa804192af581826336ad4bc2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/393.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/393.wav new file mode 100644 index 0000000000000000000000000000000000000000..9840fca367a7ab6e5e83f41515c5bea8edc2883c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/393.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3fa3cb93a717a8543710eaffb127e3ad5f5449a97d96b68920c677239880d58 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/395.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/395.wav new file mode 100644 index 0000000000000000000000000000000000000000..bab4d8afe055738cdbfa933ddca1dfba28f6cf7e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/395.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2598f3675e5f5ec245140afaf3d35b8d90824f596b1a2dc298a62e0624ab960c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/398.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/398.wav new file mode 100644 index 0000000000000000000000000000000000000000..dd667dfc400f2b4efe3b42fac1c4d2957aa02c37 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/398.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc5e79a231e076140409e01a503fe12d0dd55d4b730b515f972caf47f8a6b3bc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/399.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/399.wav new file mode 100644 index 0000000000000000000000000000000000000000..6249431601e5c8f3c1b8e998dba0dcd4c6b4f74e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/399.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b2a21853709dee71b81a52eeebe20280b3660c6ae4ac118c26212007b35df8b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/4.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/4.wav new file mode 100644 index 0000000000000000000000000000000000000000..255905a0199d41227f7ceb74de36ce6cb2381424 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/4.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08e6783a77f8fb30cfd4d132d73952c81c6250b6b7d92b1d28e668e9160cd815 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/40.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/40.wav new file mode 100644 index 0000000000000000000000000000000000000000..96fccf6f977eeca3ab310dcab41a9a8e7ac2d6ab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/40.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de51dc1fd38b9d98c7529aad2243c5f55d089f0622f7c577daf0909f77a8333f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/400.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/400.wav new file mode 100644 index 0000000000000000000000000000000000000000..f1962ba6b9e193b40198810163e002f8d7723db0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/400.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e88102ac50b0b2a78efc4ca1afdda316c2309a5d8b59fbfaaef0e0e83519957 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/401.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/401.wav new file mode 100644 index 0000000000000000000000000000000000000000..e61d61099b3da7272e459060074a7e6f107bfaba --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/401.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb9472a53b4abb14622ef1dcbfc4a574439b8d7821927ac232e36ebc5e94f4d0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/403.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/403.wav new file mode 100644 index 0000000000000000000000000000000000000000..513deeba7fdca76438f91ef840740eea4cbbc090 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/403.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddb49135e4fa42beaea31f2411f7cf8d36eca98444e036b52db70876fe820f40 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/404.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/404.wav new file mode 100644 index 0000000000000000000000000000000000000000..5048e843bc4b7a90a128ed8adbb8f0d4e6b35348 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/404.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09687b17fd1a0fa07f647d2f17d7c7575fcb5667e22eef367eca026c149f484a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/405.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/405.wav new file mode 100644 index 0000000000000000000000000000000000000000..3d7964fb367a9881614a54551fbcf85e5df6c72d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/405.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a28f77a54fc4ad2058c1d187480638d0e537a79b84c4974ca9e3045d6584b29d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/406.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/406.wav new file mode 100644 index 0000000000000000000000000000000000000000..6de378f74f4555cc6cc610fd02d6f7fa2c1c92fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/406.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d792adb63612b5abc40e84a8a9cb31788ab92f8e5b5276372b4468f023dbaa98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/409.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/409.wav new file mode 100644 index 0000000000000000000000000000000000000000..1123a2640109197b80923290e3466d35c25510b0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/409.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ecdbc1e725460ac73c5ca187285d23325136f0855dc52d77f0d4003ee3552ab +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/41.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/41.wav new file mode 100644 index 0000000000000000000000000000000000000000..ca92432c4db664878d9f959253003db91e09bca1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/41.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2383b6fa73000dc26eb20bdf19813e6b5e66f9ed195818aa2727f38d1c89843 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/412.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/412.wav new file mode 100644 index 0000000000000000000000000000000000000000..6db8c1d0eb9dd3410eb946bacf8b018af407586b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/412.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29a37afd1ae41ac72a09f08ab4029547cb138f9240ad993f5e7a944520fd1f33 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/417.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/417.wav new file mode 100644 index 0000000000000000000000000000000000000000..6c1d224a73096b62a85b77278cd53427a221d0cc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/417.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed16245fc24c0475d521c9b5990603cae50fed49b932030e1a09fa40ee26d96b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/418.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/418.wav new file mode 100644 index 0000000000000000000000000000000000000000..e02de5b1b89d427bf08610ececa66af90e37ef8b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/418.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f56999c48a862752d298d2e7994885f8cfb90f0f6c301fc3c0423a62bbe0143 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/419.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/419.wav new file mode 100644 index 0000000000000000000000000000000000000000..b810a8657a1cae18d43f6c8a6ccd512e36eff3fa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/419.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b97c023a90ddcc70754bbe400b4b9ea40d9c29699dd3d532e3de92a78fd7f3b8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/42.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/42.wav new file mode 100644 index 0000000000000000000000000000000000000000..edc8c857fedc185db84b638c82d1add0e87623fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/42.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfb875e31cf9cb1b89f62422eb7e60985071d0d2816852c5d559e7333c4da34b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/420.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/420.wav new file mode 100644 index 0000000000000000000000000000000000000000..036fb0a6fa7cc640502175dc85257dd4206b7be7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/420.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cc6eb4af67568cf7d34ec76cfb28a7ebdeab5063310985ab0e951c2d0e23c90 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/421.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/421.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd56c3e6808671c9344a8d1a12e9ebc6e2120f4f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/421.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14bd02c91f14a32e5aee921e58293f249ba47882469800575c00fcc3aa21c860 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/422.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/422.wav new file mode 100644 index 0000000000000000000000000000000000000000..ba80df1c2aa6fc1d500294b43c3d3654f199fba3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/422.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e37a160cd1eb11cef6573394f103dcd7d0d59f2cb9cc5db0a71a3348c555f25d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/423.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/423.wav new file mode 100644 index 0000000000000000000000000000000000000000..20256b92a6a5a900fb07afde431de7fde84eff43 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/423.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6e4f8063f0c89ca0464cf18682d3f0772aa59822ddc6d61cab1716c9e87ffdd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/424.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/424.wav new file mode 100644 index 0000000000000000000000000000000000000000..87c36be3ea2b32067c596ef2c7d9507c2a7a728f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/424.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b17ae131e3c22f6cecaeabeae81c0dd173cf5ae1a6e262f817904a73fce10cb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/428.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/428.wav new file mode 100644 index 0000000000000000000000000000000000000000..55455c5d9963f753cb420e78392e79e5bd1fbd2d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/428.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1068cf8919abe1f86d2886afcf911960b64cd49170540031beeef07d7f9c82e8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/429.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/429.wav new file mode 100644 index 0000000000000000000000000000000000000000..885d7c861409fcdc249df47195b53ee88547a0b6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/429.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59c71973fb612ef29e52614a47b55c4f798cf367c0e5b366e3f8d526d4902118 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/430.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/430.wav new file mode 100644 index 0000000000000000000000000000000000000000..92624c271fb598468173f5d5a094e6b82023b8a7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/430.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1f388edb93acaa867140b2e3ab5801be256beee139aa5122d004010b73712ac +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/431.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/431.wav new file mode 100644 index 0000000000000000000000000000000000000000..dceadfbcb7e28c98676326045a6a9265d9bffcfc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/431.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1afef4e83de7bb0173f3aa30e3bc2192671ee149349a9b246813b0471477eed1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/432.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/432.wav new file mode 100644 index 0000000000000000000000000000000000000000..487f8e7d7e95a3f2a9b56a82925d4a30d0e1a378 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/432.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0331c75cf83314fcd286b9bf4bdd4d93fee1e660a068152493f3928513d14cc2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/433.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/433.wav new file mode 100644 index 0000000000000000000000000000000000000000..83af8ad8ede9349d628f478e5e522467476c5b22 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/433.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb38f4006afe9360a0e531a9f7eb03fcc7e58458d33bcfc44d7025c63042e095 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/436.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/436.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b175cbbd3de3eefa23ba623f66ffa0f1a310e95 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/436.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08b15cc43bc2aed742f933564470bd7fef70b82e15aeb3fd1a243c50f6cd3e98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/437.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/437.wav new file mode 100644 index 0000000000000000000000000000000000000000..ce1104088844002f9a8fc772114f56bb0961e121 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/437.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e804011b553515b0f624b2693a584b702d9a80a1495dce77ce337be6364f7313 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/438.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/438.wav new file mode 100644 index 0000000000000000000000000000000000000000..667c0ce730f536cb2774bd24f61795bd226503a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/438.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:717b9e8eae0ad31c36d8bb045078303e503758a86616f32fa0e816f1827bd199 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/440.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/440.wav new file mode 100644 index 0000000000000000000000000000000000000000..00ee8af86a487b03e94e97e9672daae465cdb328 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/440.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c61cbf6502191703e8bea696320c2af533d60fb1493c3abd780ad9102cf0900 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/441.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/441.wav new file mode 100644 index 0000000000000000000000000000000000000000..45dea38357258381ddd32be7096a55ca0c24f014 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/441.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abe75946ce7c8e9f6845cb605cf1f8aa4476f44a638f508de41265ab474b7251 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/442.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/442.wav new file mode 100644 index 0000000000000000000000000000000000000000..94d989187d990924cc50d69c764da12adfdcfaa6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/442.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a1deb3c1f15d0abf04c41076a83c2ce9f71c3de0dca4b617cf1c4cb2c12a9b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/443.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/443.wav new file mode 100644 index 0000000000000000000000000000000000000000..0df39d3c60577e4e14cf6cf5ddb8f0543144f699 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/443.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeaec3e95533f3e73494ee9297f353b8b4f32f4dd08ca0fc90528d5c0e4e1472 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/444.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/444.wav new file mode 100644 index 0000000000000000000000000000000000000000..34e6812459adf0affc550ddc7fd2f417fbe71625 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/444.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb5230358670234debb83cc0d152da2ee28b1cbd723768b74fbe7310ff923b3f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/445.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/445.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c27d97f57f42d6078f58ba346f054eadae1644f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/445.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d60e512969ab0b56c6df9d7b7f6baa53846980afea209a382e1c917aec0d2579 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/446.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/446.wav new file mode 100644 index 0000000000000000000000000000000000000000..199abe733176ce480937152c36fc404c0393d25d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/446.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fe40f051b4042e4e6a0c97f1c64c0e31f67576e51c64a1ff8ebd7155fb64be2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/448.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/448.wav new file mode 100644 index 0000000000000000000000000000000000000000..5143569679f4d5bbb3913f50152d5e0e060fb296 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/448.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:842b8af935290a2e97d8b1b494d35144981535bb1ba3d38eb8cff7da9e7d76f9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/449.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/449.wav new file mode 100644 index 0000000000000000000000000000000000000000..ce1bdbd64490ba476e4b96a3aa453611ce4d41df --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/449.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53701cfa586ca8095cda041790c89d2b1a0dc05de958cdacc59185d1fcdffabd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/45.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/45.wav new file mode 100644 index 0000000000000000000000000000000000000000..2d7b24ada4801813d7e8ed9f4d19fdf65a5a8db6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/45.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d150240698562742d09b081eab1ef0d3a7e1ffb334c33e70f7f6b2688c0ab9df +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/451.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/451.wav new file mode 100644 index 0000000000000000000000000000000000000000..3911de3aef1c0130e800c828c189f0df082cd422 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/451.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dc3f740b98396d98042eb2c32fe23b9ff56cea0d87287ae6f80c1530923a24d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/454.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/454.wav new file mode 100644 index 0000000000000000000000000000000000000000..52ba26213326d667ab0efb33b3e9ee592fbe515a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/454.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8b19baeb746f842a9183ebbe95a8a106b0559b1e701013ff174ef443822160d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/456.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/456.wav new file mode 100644 index 0000000000000000000000000000000000000000..351e1733a225cf6e208183f54070cc44376c0b88 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/456.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b43c2e90be0e1eb101cf04b31b992874e47b0a218382ff3e201887ad0d48bf56 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/457.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/457.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d095a6a2b98e11f601ad0b0d0ed20bb8df81552 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/457.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e249e4902d9704695bbf7887d15aae6767a15f0119755f5f860d838a7331d2c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/458.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/458.wav new file mode 100644 index 0000000000000000000000000000000000000000..50858aa238ccb5c882a00d82aac88d4c135a6ac4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/458.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39365b087494240734838c036b3ba13183ff3e91b2c1c14540ae254cfa16c7a7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/459.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/459.wav new file mode 100644 index 0000000000000000000000000000000000000000..b7e01ffa2b93c4d26bb904b165caa28cba023d4f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/459.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c4d9094c0713c7fe21fcfda5397d617e9a76f704e8dfc635e1da985b5055fd1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/46.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/46.wav new file mode 100644 index 0000000000000000000000000000000000000000..42d79f8571a741c75bbd0448bc051efb3822866a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/46.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bc83abdd39a798d65ff58bb53f2a690e083d1d1b5fa146335b96fd68eb5c2cb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/461.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/461.wav new file mode 100644 index 0000000000000000000000000000000000000000..cefcd2a863eae0185805bf177cb1dbb9213ce3d3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/461.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44cf505984a066e3f828d2e268ab72edfeb062acd0e72325d55bec2074a2305f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/462.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/462.wav new file mode 100644 index 0000000000000000000000000000000000000000..6555673a30854c924322ca20138c993dde0d4491 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/462.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aacc64ddfb6f1fc2afd83e18bade3338c948d833693646b1b115c86a0a7be92f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/464.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/464.wav new file mode 100644 index 0000000000000000000000000000000000000000..32b060324ed577ee1a81e1beeee4e548c9e17db0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/464.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8830a5426d6b54333d20cccfb9219c4f1741f270c47b6071792fea6bb3fd3c05 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/465.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/465.wav new file mode 100644 index 0000000000000000000000000000000000000000..5ba688464df2961123638fd39c662c2922fed2a6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/465.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7516e4f370d6d16b3fc20e8abcdd41481ff258a2429ca4df7babd5cd5902c6cb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/466.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/466.wav new file mode 100644 index 0000000000000000000000000000000000000000..58bdbe77ae1f96ad56ee13ccb61aec661b231582 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/466.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fc8a35741d91df3ebb099e39480ad00175046fb2bafaea528e3c8c5bc07a2b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/467.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/467.wav new file mode 100644 index 0000000000000000000000000000000000000000..f333a0064959922f65d23537d8287cc2b190ee3f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/467.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37b63f3ee2693bb6daa97dd293dccbce7f5b942215d6cee48cce74e3d064e4a6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/47.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/47.wav new file mode 100644 index 0000000000000000000000000000000000000000..2ff099727b1c9d2984a3dc52ca7c523f003c48cd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/47.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54b62a54c76a8e15dbad63b8565dafd0c667670ff7c804c6b141aee2e1e31bc5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/471.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/471.wav new file mode 100644 index 0000000000000000000000000000000000000000..c5183fdbbdedbc2653a2c24a732f077780545a77 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/471.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a9153587255b12f84e90b6f3e8b13e0e76d99f866b1a164b31adc43fe5a35b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/472.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/472.wav new file mode 100644 index 0000000000000000000000000000000000000000..b6426955a4039ae46feb75c3529bd6746bf1ed73 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/472.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8c2647c5d8d0ca1a2874fe0ac94f3f64fba309a193050bc9e38e86976ae2bd6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/473.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/473.wav new file mode 100644 index 0000000000000000000000000000000000000000..8e621c446e44b3a63caf8443c78ebd37673be3f0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/473.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bd46f9eba4a4e55a921b73c28eacc5fdc93c577fa653dc1e131708575d5b049 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/478.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/478.wav new file mode 100644 index 0000000000000000000000000000000000000000..52c7c386352a706790206b4942a6678a1be5da4e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/478.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7707638058815c6311706412276e920068f4ca86563ddbb485ef2829ef706fb8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/479.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/479.wav new file mode 100644 index 0000000000000000000000000000000000000000..e8fae2f3c75505d64501e05e2c64501302949c43 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/479.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca1347dadcc51ca526f2bcfad842cf81266225e4ab66e08f130ebb70e1ddff22 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/48.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/48.wav new file mode 100644 index 0000000000000000000000000000000000000000..7b9cd45117b49b90609341946b43ed88196b3a88 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/48.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd551da9fe94e0c0387e21308e3c6ff30f3407d47f8b253a61472493240c48c7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/481.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/481.wav new file mode 100644 index 0000000000000000000000000000000000000000..c7630546cebf0bd5d0b3f3b951d08c5b1d5addd7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/481.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b9154c8f160a087b6411d0e8d2357a23264636957d7d7ba732082b97654614b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/483.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/483.wav new file mode 100644 index 0000000000000000000000000000000000000000..5201861be5ddd16426fe90c4c3b87fc8b45516ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/483.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:633d4355d870af8f546146b5cb093e1cc27f4abe3ee46789c8cfbafa3ad33bea +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/484.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/484.wav new file mode 100644 index 0000000000000000000000000000000000000000..635bb99c4b7b903005d8404f95fa7b5c5678957f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/484.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b79cb16a56ac3c8060fd75f4cce2e33d0960b81eb24516efda452ab41ced043e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/485.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/485.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e2ae7bed2cb65a9721edd99b1ffd7671853b2bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/485.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d86baf885ee1d94b7aee6d8dd6ad6d436c778f9ed018c9c380025cf40eccb48f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/486.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/486.wav new file mode 100644 index 0000000000000000000000000000000000000000..ddba11e640a249736063a9cd01ae6efdc97c9756 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/486.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2a0ee15bf1cead450a0de9fc37daa19fff0fbac21ebc8428d31df19dbaa2eb1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/487.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/487.wav new file mode 100644 index 0000000000000000000000000000000000000000..1fb43591d15bfbae08368b64122ba9d67e7cfa3a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/487.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68ef27500124adea17f7ec20ae699fa97fd7f0d4f74e35fced7fd06f60854f5e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/488.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/488.wav new file mode 100644 index 0000000000000000000000000000000000000000..e46367a62b4414779166d9296144c784f5446c38 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/488.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe7f3e9bb91ec9f09acfce47d556a3afeb2e8a3820ace7f7c79ec2801fbc5e0c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/489.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/489.wav new file mode 100644 index 0000000000000000000000000000000000000000..bec9a4befc0ea1ba59fc9808768d8eb9d71ace22 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/489.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:075d181bacdbb1261353ce337c8f27409182a8fdb38a2456038648495d9c4320 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/49.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/49.wav new file mode 100644 index 0000000000000000000000000000000000000000..063e101fa478d8c2f0e43c2006f749f378bbb50a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/49.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc86ed5ea432eb2b07dbb966c31fb3ae327677537fb2835d308cf89fd7d26b68 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/491.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/491.wav new file mode 100644 index 0000000000000000000000000000000000000000..dd3a118df2cdd424bca803c0f2f7d23ea2c91bbe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/491.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fc9722b6a8edbf7ea4ab365305364858356b3d20e82da69294c258c343a5ca6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/492.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/492.wav new file mode 100644 index 0000000000000000000000000000000000000000..87e32ef65a9caa220869d3e6bf4b1d21c6c6594d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/492.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98eeb3e6162a7feee2f1424fb0677a5e3f1ec3ba6a511222c5204964c251fb3c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/493.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/493.wav new file mode 100644 index 0000000000000000000000000000000000000000..809ef3b2fd68628bb8e8c6d5a4fe01cddd94428c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/493.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5451e883d93e80cfc3a990444d6115b99ca65ac58dea21ed8b7a59192ae546e3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/496.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/496.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e0c9432482f6e39ad94ac0146880e63ca5382b6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/496.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d51ddf327562ef0188ffbc3bdf2d535efe50c3822460bca4bfebebb247ed9c5c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/497.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/497.wav new file mode 100644 index 0000000000000000000000000000000000000000..c960e08bdef0d55cef7d9877973e811c57195d9e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/497.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22ad4757ec5a2fc22f837bccc6db63429d9b26b0e508e0b2669626445dba3151 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/498.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/498.wav new file mode 100644 index 0000000000000000000000000000000000000000..7725fc0e0395d3c61686edf34ac7950b89ca84a6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/498.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18b663b816039b045ae89a828a038b7ede4da8b27d81608b66e47e7d1860dca2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/499.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/499.wav new file mode 100644 index 0000000000000000000000000000000000000000..aefbb03f451506ad51d39b5b8208b5da0a0de3e7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/499.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfa10a005aaace31580e167432a3278713b1587b2ee2c5776e3a23f4dfc8620a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/500.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/500.wav new file mode 100644 index 0000000000000000000000000000000000000000..44f0d2c487e0998302a60525c1a2fef8284cf3ad --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/500.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:883b033478fbf100d812413e61195239545d33ca2855b1d606eade25a92af6ad +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/503.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/503.wav new file mode 100644 index 0000000000000000000000000000000000000000..c23449ec672ac1cc592bcfa22f693a1a259dd40d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/503.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce8e729ea6ca3266f7f2ae521ab75448a92f10970ac3834c97e9b747492a9978 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/507.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/507.wav new file mode 100644 index 0000000000000000000000000000000000000000..f6ab6093e7b2cab73d7bfaa4daec66d71b666a39 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/507.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:837d4de1a2033864228487c33c8303d058eb383a4255f0ca9b3dc9b5783414d4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/508.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/508.wav new file mode 100644 index 0000000000000000000000000000000000000000..5e9a37574ca6552946326220143607c9b9ae1a5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/508.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71371f88c3b387b0d9c4ee3d19ad325f467e013d56b8fa57e83a832b0e121474 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/510.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/510.wav new file mode 100644 index 0000000000000000000000000000000000000000..233d6567f303a21e707d2b1f011358593ef60c06 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/510.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f86a0784be9dd1ac2b74e04b0a92587f355ea6e40f83c4188c7fbe890144865 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/511.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/511.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a626995514235b3796d45e4e16d6b4293162af2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/511.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:769b9e5cd5a3cd4c6ec031273c545b31207e36548b4f32ac216ff43385f354ca +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/512.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/512.wav new file mode 100644 index 0000000000000000000000000000000000000000..f4f90f8df7c36e73dc88d309c7475ddd4f3fff4d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/512.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2b3fa738b8e0b0495be7e3878466a4359c7c1d1bffb5e0379ee00b79627c759 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/513.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/513.wav new file mode 100644 index 0000000000000000000000000000000000000000..e66ea1ea5445f7d39b529ace3319f43ade55a3c9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/513.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab274c0d7d9660d0eff09be25eb064a010ecd5b2ac5a37b88a3e494e38109548 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/514.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/514.wav new file mode 100644 index 0000000000000000000000000000000000000000..89a1cda40029444058b59e5fdbda371fa3cd4109 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/514.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aceb0a659597a58fa488279cb1f9b29dc2011f79daf67ddabfdda61cd0a2a115 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/515.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/515.wav new file mode 100644 index 0000000000000000000000000000000000000000..d96ac6b9db9e8a38463815d461be90a01cd3a8ef --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/515.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13c6bc39821b7cd057f967535c5288e3668e633f908ef0a9029dd2d9355f206c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/516.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/516.wav new file mode 100644 index 0000000000000000000000000000000000000000..37617d0a44ddb364348cdaef33ad8b5c2687cfa5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/516.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff91e584c129d77a7436457f347c84c4d58c7d69368182b3d14ccc485c904d09 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/517.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/517.wav new file mode 100644 index 0000000000000000000000000000000000000000..3afb840ed720590472878e52da7e1b4ff3abbb71 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/517.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45a6a4fa84996d9fed541558973d3e556ed3769506f2b4368ff59422072aa0ae +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/52.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/52.wav new file mode 100644 index 0000000000000000000000000000000000000000..58cb4a4806730177961447ddafa1fbc4ddec482f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/52.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ffd0dd1b0b90cad3e082146222cdf460fa7e156d84bce92e3c9035d3fbc8975 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/520.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/520.wav new file mode 100644 index 0000000000000000000000000000000000000000..7714d9ff091d024ef8183324113dde9e70981cb0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/520.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b297df23660ff6fe55d2c4524b40e086183db5969786861ef9e11f92b9ec12d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/523.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/523.wav new file mode 100644 index 0000000000000000000000000000000000000000..3c933800716cda85a607e08a3f9b14401c72f9af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/523.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b71315d558f8a4bdd061026c0c68d570d37ecd4b1477cacdf390b847a83fe75 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/524.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/524.wav new file mode 100644 index 0000000000000000000000000000000000000000..7f0c4e269933a1acc2b4c1496027ef4267739420 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/524.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a1b5b9acd806ba5e28373bfb3ea5b522a346f335693fd1ddce8540758a2a569 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/525.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/525.wav new file mode 100644 index 0000000000000000000000000000000000000000..d57c08d7eecb0e9431e7ed19c55c0c516585bf31 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/525.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f69c65fdf7f25193f525975bb1afa5823e5d0146fc20ffb50491456d3df7c2b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/526.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/526.wav new file mode 100644 index 0000000000000000000000000000000000000000..ba9feb5bb07090050c23db25c31a27be795b32dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/526.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:55217d8e98545df2c39326e2c36568f6f1eab92468564ba97e75c1fb8e51a371 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/527.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/527.wav new file mode 100644 index 0000000000000000000000000000000000000000..09a36de1391dedec245168bfe0031a3d93330192 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/527.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04fc59867435937929dfa92a6f156985d33ab0390cd9839a980824cfb9d00c2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/529.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/529.wav new file mode 100644 index 0000000000000000000000000000000000000000..6585be3410411bdea896edd130b45a11e4dbc26e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/529.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8be706ed56c684d41ee9b9b1864a221e4255545eea455d1ad917627affe67eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/53.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/53.wav new file mode 100644 index 0000000000000000000000000000000000000000..0b2849337e3a886074e4efcf28f306aba70cfa3c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/53.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9662ba4bbc2f4ae62597e9e18579c33ba345e9820f2ca5e67eb32f03d096ca56 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/530.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/530.wav new file mode 100644 index 0000000000000000000000000000000000000000..49639bfaa9045df94ca1f7920416a2ac5967d667 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/530.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:372e390e5170f1586b89b615f266eb86d38372b332ded869be0a5419b5c2d813 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/532.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/532.wav new file mode 100644 index 0000000000000000000000000000000000000000..4612733b045cb2dfd31c5bc0b38906677b55c39f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/532.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:890edd9081fd02ce5c649e396cccfbd1de2ee48cc8307613ccb363ff631cb822 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/535.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/535.wav new file mode 100644 index 0000000000000000000000000000000000000000..57b5786c20e760e165a1d1d613083f4c3cee93c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/535.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e24f3961498f8eddb0549d0512457077df50d51ff60a251c048447c586bef21 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/536.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/536.wav new file mode 100644 index 0000000000000000000000000000000000000000..bd0029c53dcdef6431ae8885d205b32ed1c840e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/536.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e095ee4b9fc3567cffdfd81ea8734e2aa944e5afbf403e9b969014d6eed3fa47 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/538.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/538.wav new file mode 100644 index 0000000000000000000000000000000000000000..b3457f15aa356e83f097bbe7545931037d2dcf11 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/538.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c63994515d63cb646459b866fb86fcec0e4edf64e473380a86185bd41b2f660 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/539.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/539.wav new file mode 100644 index 0000000000000000000000000000000000000000..3bcb60108d48fd02c2dd44a8edf4e9197a0402bb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/539.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdd82071b477269dfb39fd836582fdd0a54f159cd16d16d828a23cda77da2301 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/543.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/543.wav new file mode 100644 index 0000000000000000000000000000000000000000..9ce80b4af5c5a61d068e26ab273f1514a2ef0078 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/543.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7475821548e833ca85989edd3522584e656b7b9dc2cd9bfc1f0f139258e172dc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/546.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/546.wav new file mode 100644 index 0000000000000000000000000000000000000000..d40c543f997efda9092d27de76903a23403073b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/546.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb9ea92e02f0859691b64243dedf5dd51759bd12b0f9d0a836f5838ac4e4ca87 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/547.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/547.wav new file mode 100644 index 0000000000000000000000000000000000000000..b2679c6b3dca0888f2173090992f1214dfde2ac4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/547.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8deafb281c63aa6df5fef186903d2dc203dc4b35a3b012d693c2103f4df13d1d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/548.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/548.wav new file mode 100644 index 0000000000000000000000000000000000000000..146eed349ad719981823ca9db5dfff9a00fbcf84 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/548.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:386bf3ba00bbb850a03f722a9947e105fa5d15857d1e1fb36becc901af316b73 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/551.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/551.wav new file mode 100644 index 0000000000000000000000000000000000000000..631b884e5889acbd1221a5abe4d587df10ef18f6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/551.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ded64965c0e76688cef658051a1127b440e9231ae4def9619baee014f4bbe0b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/552.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/552.wav new file mode 100644 index 0000000000000000000000000000000000000000..d461e53c40a9916ca84600bf3c771e88d12a6133 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/552.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db4272d2c19fd19a8d3853d146e8e71c655f3a19edca1dcdeb06d2eeb22fc71f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/553.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/553.wav new file mode 100644 index 0000000000000000000000000000000000000000..3b215150e73cb52f4a8b9a29053e0610c25a3d95 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/553.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ff8738c9a92bd51751d75090a9660b1c6978c954020a93602c4e919b3d3d333 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/554.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/554.wav new file mode 100644 index 0000000000000000000000000000000000000000..0ade641739b99a1c74310dfb84138225df2054e7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/554.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d4256464b3fc142d4ef289ec071b92ca2df6bd48ce0e6dbc9d3b132f8482227 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/555.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/555.wav new file mode 100644 index 0000000000000000000000000000000000000000..d5f9ca680152ab70766eca9801e999112f04438a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/555.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7384b6b6d06a9aeed6e38bd5e307ce986ab34a11805aee55e51e2eb71908041 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/557.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/557.wav new file mode 100644 index 0000000000000000000000000000000000000000..c0d636e745c985cca775e53f93a68316130b201f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/557.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4792e47e4352d99a33cf693c4669ff67b682fadda28dd5ec20110b23e5d80034 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/558.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/558.wav new file mode 100644 index 0000000000000000000000000000000000000000..cacc4b2f78bb473c4d3015b531b956592a9513fa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/558.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3cc2fee054c7e0ea36ad1c8a74922d194f922fb03aa44a93c715d81c7cbcd6a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/559.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/559.wav new file mode 100644 index 0000000000000000000000000000000000000000..e138e9536a413543cd9f4be00056e69c9b2a8dea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/559.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41954a90bead228d487d53636348b815345d980a437abdd2f578be0960c51049 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/56.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/56.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c6f5f9a6cf5c5f2d4f261ea3eace94bf8b1d10e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/56.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a99c1a19dc4ba6a61e9b7d53b5f5ade2f5c335cb4db33c945f85ea62d14c385e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/560.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/560.wav new file mode 100644 index 0000000000000000000000000000000000000000..dafdd6c270f7bfc28d08c6abe712e6578b544563 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/560.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:919953eee3e8a9d17793dfb972ec1a61b723bc1d7f813b2cbf9e9dd251fc8674 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/561.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/561.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebf026e1270327a80d052f29295f57dbcd1ebeac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/561.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:265c36986d9c24a15d4f6c1c2e207ea34a340f9582b9c66dcac7993413cdc214 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/562.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/562.wav new file mode 100644 index 0000000000000000000000000000000000000000..0071dc5fe21cdabbe46443eb14d1ba2f07b73c2c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/562.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29897d80f6c01e938511c00bdce008ca2a8a230fc61fb42cfe71fae1ad1a21d7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/564.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/564.wav new file mode 100644 index 0000000000000000000000000000000000000000..894b259561da07194a08df5fa81743c146d4c73e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/564.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0c540dc5608a136a16ade0f1cf8b6ceaa24860327e528f21be491dc236d56c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/565.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/565.wav new file mode 100644 index 0000000000000000000000000000000000000000..48f8e830294dba1af011d177c61d0b46f7ae7589 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/565.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dce9db6b80bf27297985528841e2bdfedd8433cb7925d90fcab58328179d5fee +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/567.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/567.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bf945ca963285bc4263c72b5add12e3123e9f8f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/567.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abf3c967ae4267af8b75fb6eeed2bcede0a6c0397c03b1e35ee06317b6df8370 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/568.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/568.wav new file mode 100644 index 0000000000000000000000000000000000000000..8bd2eafea9df405685031389973f3870800838e6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/568.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d66f61eec943959b313d213a1bbc52fc35ce505198e9cd5b0eda3d3f51e79963 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/569.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/569.wav new file mode 100644 index 0000000000000000000000000000000000000000..a44eb0a80d6ab60aeed8063e80c4e95f96594ca0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/569.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79faea6089af3e1f49b9f8141631ec88ce8d23dd623acade90e512dc0109f00a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/57.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/57.wav new file mode 100644 index 0000000000000000000000000000000000000000..b56e98326624f3d5a80a26a6f7249ef23a0efba6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/57.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a16f191418cf787cd250ffdc82613ddba57def72d8f243ae26b28d0d16325e0f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/570.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/570.wav new file mode 100644 index 0000000000000000000000000000000000000000..70274379d0fe4f6286cb2b721fc761030e5ced0c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/570.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:377a579f99f853bc383a84ab05effe2447ea7af82d9d1626533623e67ea3e089 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/573.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/573.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe3ac3e3c5b714108680cc2960f57201b7752816 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/573.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2549fbddf49c3eafb008d9c6fd35055cccb39139a5d02f6ba21d67312fe22d1a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/575.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/575.wav new file mode 100644 index 0000000000000000000000000000000000000000..4397a37746e1afa11c4350bc6fecbaabe714e3f5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/575.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fa4fcb6e3965bb2aba3b0015b8bec3aba76a719da4986f01cf17e2cf1e6fe67 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/577.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/577.wav new file mode 100644 index 0000000000000000000000000000000000000000..56248c69719c1b8e800a8750c8711d2e0c6cea52 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/577.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f04901d015739e9e57152e5d865572e301866d531f1e7ee640fcf746266d2d4d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/578.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/578.wav new file mode 100644 index 0000000000000000000000000000000000000000..b77a41aa9124c04f18c46dd16c1a6559c52b8014 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/578.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:484f716b4de2c48a6c08d622094bcde17f0fea1eb4ef5dc1f9070ad95e5c1612 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/579.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/579.wav new file mode 100644 index 0000000000000000000000000000000000000000..48ceb9d91fdd2977c3529f9e919ede88f4cb2d83 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/579.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1da0a3e970ef3711358d0834a5c353076eaeca0650380b9bb5770ecd18bfff9c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/58.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/58.wav new file mode 100644 index 0000000000000000000000000000000000000000..441461f89686e4a2a2015171502d9cad915af0e7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/58.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8c5966266e019d1523eb7085823d9a264018c18c6989c568c4882c9e10279b3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/580.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/580.wav new file mode 100644 index 0000000000000000000000000000000000000000..9c620ceafa5ad277dff35436a21b176b9e1e5339 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/580.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29f7559c38c11fd65bd53716e02482b38bb715f583c09e283db4c3ce83c57b8e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/587.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/587.wav new file mode 100644 index 0000000000000000000000000000000000000000..3023e4de603cb5f7d0c77fb1f9c5bb949731b7e9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/587.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08b38fcffbdd678b9c15f012156743e6b3721f61571d9c82b5f9fdbe6a18b6b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/588.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/588.wav new file mode 100644 index 0000000000000000000000000000000000000000..5542cb34d343077ff74ed41392819c7f5eed6398 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/588.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df997a2180dd646695b3323579878b74df8cb2ffd9e2b651774eb029cd4d74af +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/59.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/59.wav new file mode 100644 index 0000000000000000000000000000000000000000..db863c7a0351ad14e47c93bc0c7ff8f95a20448d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/59.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16ae523bd9078bd06b8ad5629919223e3758efb13551934da0f6427fe98e8c94 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/590.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/590.wav new file mode 100644 index 0000000000000000000000000000000000000000..9f9ae7a1979bb6c3e11c32e62d2214afa3023781 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/590.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20b2322aa390604d85942ee0c109c5d7cc3754eadb3143dbe672cff23766d681 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/591.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/591.wav new file mode 100644 index 0000000000000000000000000000000000000000..9aa912b7816872612b8e71c101896a01361c1767 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/591.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d684e7c018b816bd37123a5646c671c65b411c83f1ee628ef7aa0eeb4a187de1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/592.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/592.wav new file mode 100644 index 0000000000000000000000000000000000000000..fe7c5c99cfd8da78d101032b6f910e589b622f09 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/592.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f00729e0979040fe4976bf8b436047e2e88017674f7f00b5aa1473d704bf0e8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/595.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/595.wav new file mode 100644 index 0000000000000000000000000000000000000000..41b286e4c61fed69eaa4dbf6bac2cc597e17dfab --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/595.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:141c9f9bf58c027942c49e7846210efe8df7608c3f62e3a8ed07798cc6420d2f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/597.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/597.wav new file mode 100644 index 0000000000000000000000000000000000000000..5118566621e536ef2257609003d95594a064f55c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/597.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4a4cb94b699e3525783656d2f529fbe5e4926f5dc3c9454d9a4940f67e9df8e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/598.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/598.wav new file mode 100644 index 0000000000000000000000000000000000000000..0e3f0b33d5e9164b2ecccf1e49790fc06f055ed8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/598.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8d9e4c50848828864c16fcb87e08bdf6432824876599a9f987d0157d13c5562 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/599.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/599.wav new file mode 100644 index 0000000000000000000000000000000000000000..250f81f674a841b7fe194b29a5d17fb924587b75 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/599.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6453237a8fa6fed7dc92ae1f2e3465ef416c8bb8b53b31597bd9c2832ca2c1a4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/6.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/6.wav new file mode 100644 index 0000000000000000000000000000000000000000..98f1e01f880f620456f6d196a31f3813cde42730 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/6.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ecb495288edcd0e33e1f29a96540589cd7ccb2506e93d23f04c0cbfa33a9734 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/60.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/60.wav new file mode 100644 index 0000000000000000000000000000000000000000..72540861f41e30b0091b6a282ff158095a6bd48a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/60.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9b1bf641f0b50fc2110f14545ebce711490b5e1fdcbe7c648b486e426fb5d9a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/601.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/601.wav new file mode 100644 index 0000000000000000000000000000000000000000..2cd51c7e78ff8f3325ea3fd3c0020bd2f70d5164 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/601.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4590b8fd0e95fb925c924d355e61284ba443d7e38f8c5444d828205d33e13942 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/602.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/602.wav new file mode 100644 index 0000000000000000000000000000000000000000..e76ef06610e922f91d8a2bff4689308db6d19f61 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/602.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93c3d54073633271240f697522a54eeeb1d5f1f6fdc82a98ac8ba4c6835bf3df +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/605.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/605.wav new file mode 100644 index 0000000000000000000000000000000000000000..4f8ba676055603b8f63057993130066563764d8f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/605.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da7a55de84d825266a36d1f696010d71641a23c35ef08ef8a5fa6a6e728300a0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/608.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/608.wav new file mode 100644 index 0000000000000000000000000000000000000000..fa4b2f6a431c1a61441414144a11ece230ab1f45 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/608.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b314f5733fa55048b3ec10357e66961864eeeb19335639d814949db14dd92e14 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/609.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/609.wav new file mode 100644 index 0000000000000000000000000000000000000000..d02c04ff9a2ad1558dabfaa69afa2860afdda338 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/609.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d88c0fc246e363506f78c3eac60fd7061ea04ed6ad808e04462c2a6717ff84b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/61.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/61.wav new file mode 100644 index 0000000000000000000000000000000000000000..3b45a82b5f949ddb4a880be613d21f26ca92766a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/61.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a182164307ea068331df97b0398bff0f82f9aa41d72bc9f702ef8fd97b263f2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/610.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/610.wav new file mode 100644 index 0000000000000000000000000000000000000000..313a016b295ec7219368119b7403b83340a47e6d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/610.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afa5dd7a35fec83ccf2c4da36951a6459b92366c063d880df623037cf80da58a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/611.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/611.wav new file mode 100644 index 0000000000000000000000000000000000000000..5761e5e56e3dd42c4bf37dd526a8114cef73637f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/611.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32f89119546f40b70b6d72625db0bac016f97c7f2d07a054f6d2d46bb5f28807 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/613.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/613.wav new file mode 100644 index 0000000000000000000000000000000000000000..55088552b2e6a07df0379d66115a06813d44e22a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/613.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1868ad6687cd2d0adcc31c904a322036974a5006e6b434d8858d6cbbf8c4dfda +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/615.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/615.wav new file mode 100644 index 0000000000000000000000000000000000000000..c6efaec5d7bdd8b7709bc67a2697f602e54f9c07 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/615.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f200970916fba18512c94772b55e0102d0bed5070cb7d2791b951bdcbad3ebe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/616.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/616.wav new file mode 100644 index 0000000000000000000000000000000000000000..fbd2aac8a602e15d3db4d07b7402f304cec82cbd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/616.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b08dfb633151ac6868bf422da26b4d1110a224fbf80af80e4c04275e1151d25 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/617.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/617.wav new file mode 100644 index 0000000000000000000000000000000000000000..b6932a974684b6bc7490cd7a9a8ba5b4fdc23133 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/617.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7979657be2716b01141a53e7e178f7b1d628e34d1461ab618eeea3f129b0250 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/62.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/62.wav new file mode 100644 index 0000000000000000000000000000000000000000..cb90cc1325a0562e827cb5664d9778af12fe4bdf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/62.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae49de99f62f7ccc13d4f2332cd0c417fee135c63f34bf34db04a154e93ce157 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/620.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/620.wav new file mode 100644 index 0000000000000000000000000000000000000000..8163e085056f18c97cf2be62797a223cf7cc0518 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/620.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:970bdb4732d9d6d4078852a61b58b1c3acd7224ac921d8e6eb83c925120bacbb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/621.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/621.wav new file mode 100644 index 0000000000000000000000000000000000000000..72bfd34b0b7605eff4a2fa8db7a610eb75f39574 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/621.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19aa91be2ddc5e76dc45debc00948278c3ca05b8632be08c5fac63d34c9d684d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/622.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/622.wav new file mode 100644 index 0000000000000000000000000000000000000000..39b29c207c0737021762b66151ee0cb60bb623a1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/622.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:faa2a1b5486ba1a46a0778b7f465c50a8ee82152511e266ed208d5d4b72d63fe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/623.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/623.wav new file mode 100644 index 0000000000000000000000000000000000000000..9da600a479ba7821fdc98846d0276260be0f3637 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/623.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df1da3c96d8c6f81a66b2ec00e228580b080ab9d4aedfa0e5790762c89d05e44 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/624.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/624.wav new file mode 100644 index 0000000000000000000000000000000000000000..9bb71ab55f89638a307491d5792765e5714dd941 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/624.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3328fe7cba6a21edc8f80e4b290c16721175aa5398a83a16d330a09cedc4dfec +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/627.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/627.wav new file mode 100644 index 0000000000000000000000000000000000000000..375fbe22b4afcc93afe8daf6628314dbc213e5af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/627.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cccf7b476e9b53e065a163ac8d83645e94db2d4086affafb5ba58d5247b9dc98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/628.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/628.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a1a7ca2a27fffa7325febff01f55080e5874910 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/628.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a1f2b29c131dacb0a35cf9658eb7a566e0b6cd11f9b34a3906b8732103679c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/629.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/629.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d49a9666100fac9fdd06d01e36cb82b31b25b6b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/629.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d963e7f96ac3d54d0fe9b94fa56d2d56c04f8003b9714b1f9bf5b9f3e90127ec +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/63.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/63.wav new file mode 100644 index 0000000000000000000000000000000000000000..85961001ad9c230af871a41a320dbc2ce728b0c5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/63.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d1e32715f2d0624de6e0946a7ab4bf0101be359b4952c2bc11d09957cb6f4a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/631.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/631.wav new file mode 100644 index 0000000000000000000000000000000000000000..d1b7e3537bd19d60e9f16788e6c84ceecb6f5577 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/631.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34bf3c1949fe687477dc0da68622472d56fa29db5af11e67eba3c3340b4d129d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/633.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/633.wav new file mode 100644 index 0000000000000000000000000000000000000000..49f82c727276e57453f0af8a9983c90cea7fc4c6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/633.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a586446d0bd9f3ccbe1539e365a5b301357a23eb1b46f7a232c9d4740517bf1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/634.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/634.wav new file mode 100644 index 0000000000000000000000000000000000000000..33b9c014cc99115b20be8b5fa80d62da2ab6fa55 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/634.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3e8906ba29ca5d2dca9d1dc66d6f2e25829a4353378a4300e51e4c607d9f8b4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/635.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/635.wav new file mode 100644 index 0000000000000000000000000000000000000000..e4db268b5f9d648eacc0e2b134bd2f82ea666de4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/635.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddc6b8bbea6f0d6336488cad4a7f7c4d91cb2f9a5434d043a98b5fc724586d0e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/636.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/636.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdc8a07c445ccfe62f118e110ed1b8f990fb404a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/636.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:323a19cfc8194c3c4a12a839a5711adbee5ad4bec3e6d6d8544c16e98060abcf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/637.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/637.wav new file mode 100644 index 0000000000000000000000000000000000000000..8fe72acef27927896173643e64002e9c8b0640b3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/637.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2173ad21e0847555cae4eb13d832bf2087aa42f44439ead6ca9e475b81458eb1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/639.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/639.wav new file mode 100644 index 0000000000000000000000000000000000000000..fc95615226196f9d11cccebfd55a492f90b0f6d8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/639.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3509ed0267e8e8fc70e898a961e7eb7561b319df3afd8e37cdfe85c8e530e201 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/64.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/64.wav new file mode 100644 index 0000000000000000000000000000000000000000..625f9102d1ffad517c3245b4f442bb06253003e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/64.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8363e8ce61acd0d5cbb30109ccca3d1489d1d0a584fb4e87b274e2aa169fb6c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/640.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/640.wav new file mode 100644 index 0000000000000000000000000000000000000000..5c11391a4a4053cbf042b307dab6930b8e4777db --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/640.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fd5b3ddb99ae7a35c9caced2c3333795d92254f7ea812e6ec011731e20a0d05 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/644.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/644.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d7a8f568eccee84a65c110eade573d2768949d1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/644.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:925c861b95e0e1e1d11b3dab90335c07a124220ac0f1ef43e8510299c4c72cdb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/645.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/645.wav new file mode 100644 index 0000000000000000000000000000000000000000..d704d98f14130b6093f69a26b6cd58a477df3687 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/645.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b0b310a125fe9cb4c33dd6e9f540bed8ee6fce7a0bad973c6073670e2ff614d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/646.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/646.wav new file mode 100644 index 0000000000000000000000000000000000000000..249d88bbde4f773735987ccffff7cbd3aa1b6440 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/646.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d92cbf0686f28fa54fa7a0ac8cf2026aec56b4a934bdfd86b2e73cf51b7ef99b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/647.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/647.wav new file mode 100644 index 0000000000000000000000000000000000000000..2573311c54f6cb8ee3c4c67b49e4e802c5426dd0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/647.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1652406be1027f6dbf6228101352f31ff17f4090a4a8dcfe2a69b0364bf04e71 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/648.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/648.wav new file mode 100644 index 0000000000000000000000000000000000000000..441b7ed7a45d9a55da3fd3e79b678d868c46b8b7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/648.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0ac1ca145b32b919fd55d7b387f7fb90437af733c9cfd5517c37b7a9427985a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/649.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/649.wav new file mode 100644 index 0000000000000000000000000000000000000000..45327dfc3c924b4ad3432263550f0ebeeda5cac2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/649.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28f26397c183d48efa4c2fa55b3bc82d9791c9ea561cb8b59ac6c44f978248ae +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/65.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/65.wav new file mode 100644 index 0000000000000000000000000000000000000000..1cc8dd5f44b352935b8dbde6c4c1b614855f6fc7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/65.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:148d2e2f8f2cc5667c01c944f879c19d696cbed6d019faf9a4da78981b544eac +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/652.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/652.wav new file mode 100644 index 0000000000000000000000000000000000000000..ba5ce7cfc35493fcadeddf3b2f4c4e74985d2d66 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/652.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c3801511b68a5203258e0e21ce62639818caf5ab6a0ad4d0e32aabff557bf40 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/653.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/653.wav new file mode 100644 index 0000000000000000000000000000000000000000..95fc45ea5c9ee8327f72c932fbfaea84cc52e24b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/653.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef77e27a73c8af0c48e0c6c3e55889e6b22c02177bcbe1f211d16c5ba0425b79 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/656.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/656.wav new file mode 100644 index 0000000000000000000000000000000000000000..480aadc0e003a9726e183260c50c034fde3d217b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/656.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d217445585a0dc79c1a6dd5bd71e3bcbacdb9c0a2375dfa3454dbb5f6969d1bd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/657.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/657.wav new file mode 100644 index 0000000000000000000000000000000000000000..20a963b5a70261b4062dbcbf4dea59c57f0764b9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/657.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb2484e81fddcd8ea4e2f449f25852548c26517142ac4b78aaf48ad2d6c7bb57 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/66.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/66.wav new file mode 100644 index 0000000000000000000000000000000000000000..aa90a7fddc7996d9b815113e499c35ed9ea6ea81 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/66.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73949aee3a776bf14e9abc80a0bbb20c58b563d36e0d703d8d5c25ecaf8f3d04 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/660.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/660.wav new file mode 100644 index 0000000000000000000000000000000000000000..f5927952cee390f9cbd08103cdcb8aaed0e85184 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/660.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30fc34ddf836bf059f79d7a6789db6a78fcc18eef516ac090b159cfc7b2462b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/661.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/661.wav new file mode 100644 index 0000000000000000000000000000000000000000..f12c6a53e1e5d3e7cac3a886f99a3d3d147a7a10 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/661.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:feb06ad4861fd6c394ee3fd5ac6041f8a4a9267b18cb7daa2092da23478c2537 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/663.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/663.wav new file mode 100644 index 0000000000000000000000000000000000000000..85724368796b9f5fffc1c4eb7805ef7a842dbd6a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/663.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d8b344038488fc89fd611baa739137fc1c3a096167b0c11f17b9a4e6cbcc168 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/665.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/665.wav new file mode 100644 index 0000000000000000000000000000000000000000..327d7f2036ccb390e79abbad94d50278523f6de2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/665.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:acee7285419cb0a43dde5f039744dab8bb27c80340f7baa449ecdbe582d7549b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/667.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/667.wav new file mode 100644 index 0000000000000000000000000000000000000000..c1372f838c90c810b21c04fe888342fe53a1b6d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/667.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:405d45701d82e19f02d876bebae2dce4832bf3a0cb54b958a454b63169873b72 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/67.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/67.wav new file mode 100644 index 0000000000000000000000000000000000000000..72b75107aba2ca73820ca805dac47f875648e650 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/67.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d550bb98474e36601148289bd7bda58c434513f65b74a2d5b90fd4fbddb56960 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/670.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/670.wav new file mode 100644 index 0000000000000000000000000000000000000000..cedfa5f58b8f4b885a49e827af8882a8b149ad87 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/670.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a3e1a5ceab9c87711fdf1c2b604dc519aae32ed00efbebdf6106525e823d957 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/673.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/673.wav new file mode 100644 index 0000000000000000000000000000000000000000..40b654d77f312f7ee4c1fbea43178f178abac3b8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/673.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:059bc680db4b2cf36656fa597761b2a23950f4ad35c92fd10b5e9e1761b5856d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/674.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/674.wav new file mode 100644 index 0000000000000000000000000000000000000000..d706536f7b2612d10341abdd1a1e260e6c3eccb5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/674.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99759840ba9488fa6fb39bb4d7af97484a23442e4ce95cb9627798e1aa8e33fc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/675.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/675.wav new file mode 100644 index 0000000000000000000000000000000000000000..f2293c33c5402a4d7f3bfbd859d6869c1fed8329 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/675.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbfb219b8996f04f50d05b6ee4e74eb16a0ffb08e702f1269642536cabd05c16 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/676.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/676.wav new file mode 100644 index 0000000000000000000000000000000000000000..6cb92548164043655b418c2cb6e076e9c75c99af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/676.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e42bd4f0ce7f0778e23027841a97eb1547d8a7dee9b5464bf438364d40c07777 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/679.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/679.wav new file mode 100644 index 0000000000000000000000000000000000000000..9672adc04bd27b98475089904c6f9c5f8116eda4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/679.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c2c611f66dd936b899ebb8d2a05eb0400e356647538e50ff10bdb4791ab4ab4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/68.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/68.wav new file mode 100644 index 0000000000000000000000000000000000000000..a87335383914c027c8a703e59770a65d941d74a3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/68.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e9a58a1fcac86abe2806bbeff53b66fb688432ae1b4d2e92677aa190b60a2ff +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/681.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/681.wav new file mode 100644 index 0000000000000000000000000000000000000000..cc773553c434b2c7d4694003ebf6eea6fcabae56 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/681.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71d93330d9abc47fa3d67bc5a10659342f24b9b166683ef24e96bc93f2066f26 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/682.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/682.wav new file mode 100644 index 0000000000000000000000000000000000000000..bcbdc2f1c65f3a6db17e30794b71154c12c0014c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/682.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f258e99c3b778233c94390481a007f0dbeb00144703232abd7f49e47adfc8438 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/687.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/687.wav new file mode 100644 index 0000000000000000000000000000000000000000..4f0b504d440dd7a7c0af8289eb6a96acf1372239 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/687.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d7ce256476d8c79390d2d33ae1dccf76db2a50512dbcf168fe6c542e14b7eff9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/688.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/688.wav new file mode 100644 index 0000000000000000000000000000000000000000..a88093dd9a8c10680822dabaf86e5c55695f30e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/688.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f22a44707c2bc371cc82b9b989b0336240617adaac9fe328a6fca3d181cd8e4a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/69.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/69.wav new file mode 100644 index 0000000000000000000000000000000000000000..ad5000014ba4037aba9eafa143e555c35cdbd089 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/69.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:474924709d5cd91bb1e98a37e7c0886c9c50c9f0eedfa7fb67157b278f1e740f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/692.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/692.wav new file mode 100644 index 0000000000000000000000000000000000000000..948fa07d2547522e48c38fd4f5312d88ebf5d1e1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/692.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ea0026d081cbd6547d4dc28570ea9a8502b77fae8b353328259b1d058f0a1ff +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/693.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/693.wav new file mode 100644 index 0000000000000000000000000000000000000000..4f27018305a3888922bfd92d937779a143628d43 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/693.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92e7244abd407b25062d4f921c0552a51973ba0f8a52ff9e112b92d0fcbc92bd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/694.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/694.wav new file mode 100644 index 0000000000000000000000000000000000000000..e08e6e879e5bee03952ef01b5f2a504208a59b1c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/694.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56f5c686b10e817a0e0b0b2c395c0820b31a7427115d7f58989389e1877340dd +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/696.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/696.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e48688e33833a6a24c5f2ae8892af0023b7d7ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/696.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b33e5efa0d42c12948e76a08c0ab99a5ae81a3e3ab8f28cc524c26592ead971d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/697.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/697.wav new file mode 100644 index 0000000000000000000000000000000000000000..f8d209d483427842ec76c66b4736a1f53be9f425 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/697.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d978821a8b861bdae99fc3e3606f5d1b0c3b68bc0b62da14e68fdce57f882b4f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/699.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/699.wav new file mode 100644 index 0000000000000000000000000000000000000000..af3c0c6deb81ea9500cfa11eb77d28a242d466e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/699.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9adc3ea6eb4e2c4114c25dd69dc824a6904155fd050679d978f4ea40313c6a3c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/7.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/7.wav new file mode 100644 index 0000000000000000000000000000000000000000..14d86a1c510695dabf553bfb01ce9b69c8c477ea --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/7.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19ded7051f546baab32cdb3e0fe333b4a603dd5f1b93016b5516a16ee071f600 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/70.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/70.wav new file mode 100644 index 0000000000000000000000000000000000000000..164efa5685e80847c10bd0b22730b0c57bf57f16 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/70.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5747a4e8d445acdb48650bb97ff3505a2d3ea59eb4c53b395d42c144ef3a0abf +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/700.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/700.wav new file mode 100644 index 0000000000000000000000000000000000000000..1368c5bfd96768320105407702e56a20da58ecd2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/700.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18399676b369e3cbe26b44644335b49761dcf1a0ebb278f253d556a8c24dbe0c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/702.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/702.wav new file mode 100644 index 0000000000000000000000000000000000000000..b4090b99558ea5a6e9f4fba9e779de7117a4185b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/702.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba2e904b11eca1b194d6176f2d2f929d0461e2f17018d4a50b53c5ded7391c7f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/704.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/704.wav new file mode 100644 index 0000000000000000000000000000000000000000..79269c34ca29e4197025463be6924959e52c8f24 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/704.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1d172d80c52c4057002473d131b44b5badff674387011339f3334dfe6b94941 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/705.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/705.wav new file mode 100644 index 0000000000000000000000000000000000000000..48bad1a4b87ca6f32727e5bcbed37e2271a0a1c1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/705.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:431d61dbc14e995daf4908b11a9f36c1bf649c9c63bcf1e1190ab6688cc7e44e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/708.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/708.wav new file mode 100644 index 0000000000000000000000000000000000000000..2c5cffd98072e9abb1156f6b307c2ffec5abdcc2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/708.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39478ffc02c09e6acd955e84324462d3aee7862b2ffdcff16bc05446d31dcd6e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/709.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/709.wav new file mode 100644 index 0000000000000000000000000000000000000000..db0a628be22bee9fa60f4a7d79f2e8bf2d671c48 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/709.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:329f855c9315903623a1ed30b9f1a1b710861cea111f210da909d097062f64d9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/711.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/711.wav new file mode 100644 index 0000000000000000000000000000000000000000..ae291c4ad1a8f3368adb84c391961cfb210b5ec2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/711.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a21e4c38c793e1cbf032497fce8bbdc147d882e0aea30d1872679a6318d2a60 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/712.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/712.wav new file mode 100644 index 0000000000000000000000000000000000000000..7eabfba6fb871ed65dac2abc1c214fa67af95b26 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/712.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f62604ede9f80dd104d0c72efd6400225ea1701baa25327ce34a833ea8763bb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/716.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/716.wav new file mode 100644 index 0000000000000000000000000000000000000000..7b8f7e7ecba8f6e8b81e138aeafb40c3a496f560 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/716.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c44acd749d027c83ffd8a5535b7e663440470ebee27db0e2778d72f095acde2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/718.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/718.wav new file mode 100644 index 0000000000000000000000000000000000000000..8cd4a1d87414ce2bd1df6e28b109e9da57d86dd1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/718.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb1f56b615eaeb75793618e14fb92cc48cf4f23bb86b1736329d7de8fe2d1532 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/72.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/72.wav new file mode 100644 index 0000000000000000000000000000000000000000..479fae8c0725e633daef1c68931cf5c38edad220 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/72.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bd246f2a0e5cd575264b802140b4ba6391f44557fcf6f23aa0bc62f9e5320b9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/720.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/720.wav new file mode 100644 index 0000000000000000000000000000000000000000..d9250714e964f536be1550ba48b49a18da55c4e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/720.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dc2f272ffd428074a2a1e68ac0f178bca7d5c8cbfeb885b0f5e22d4f9b237b4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/721.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/721.wav new file mode 100644 index 0000000000000000000000000000000000000000..06ea4ce71c181a374a7f5002f6e5cab56185a036 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/721.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f840ba41ffac926297ddb5604e7859a9cb3ec6a3bc1ecdafc6b2475446d3c808 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/722.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/722.wav new file mode 100644 index 0000000000000000000000000000000000000000..a44264ff52de8bf7a1a51325cfc380164af6bdb8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/722.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5b5c369855f67fcb8aa9ffe4ff41b086491ad0fea1175b5d35d82eb66a3603a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/725.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/725.wav new file mode 100644 index 0000000000000000000000000000000000000000..91009b9b06ca20e653e047f5341dbb629468df94 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/725.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1af29fd3f8071385beedd8369590bf4ff976e8fb56cbfddbcde1c31bef662244 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/726.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/726.wav new file mode 100644 index 0000000000000000000000000000000000000000..71985a883ce36529c8630c3f6a81e277b0e41693 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/726.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2794ed21e032d585311f5c12844d28c5dd1ef132e10be36e3602ac027f2f4e85 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/728.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/728.wav new file mode 100644 index 0000000000000000000000000000000000000000..79a0db108bd47c016bf448f8fcf7366c6fbb0784 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/728.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06efe03104456556e6984d3f8a1735a0f114f488156c5db41f812bbaa1ed46df +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/729.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/729.wav new file mode 100644 index 0000000000000000000000000000000000000000..9479822b41e34fa96519dea5e464e4d3b711445c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/729.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e0969d41c09c3d92786549dd63e60061e93dcd75db2921769643f8b35203a2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/731.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/731.wav new file mode 100644 index 0000000000000000000000000000000000000000..4d832d8c2280b2c7ba576ed77218b97a7f838efb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/731.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:406347946dea3a21d12cb8f46489633d80e29aa14030325e46769693f0c54bac +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/732.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/732.wav new file mode 100644 index 0000000000000000000000000000000000000000..199d5eed7aaccf42ea2293c25ddce3da78575a37 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/732.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14683c5d3bf15f14df91ab760a0889dbd5e543c2c2b8e92885bce4277a4856f0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/733.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/733.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef96b2343e723704af378965d59febb926f6dd95 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/733.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b8297a65e6ad428388621ad408ff64b35026983f73dcf41e2aa944aee3fcec8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/734.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/734.wav new file mode 100644 index 0000000000000000000000000000000000000000..e143052269cd10c94d0e2bc9ea8f06e9a63ad710 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/734.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a7ba20a2daaeef0bc5b1178e4ba9986021cc9cf755b09976d91f1d4d6b49436 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/735.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/735.wav new file mode 100644 index 0000000000000000000000000000000000000000..344fca835391f6142113442f30f4e3d21eaf18ca --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/735.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d5273c151b91db2751cbb1ad11685c03ea01e904bb61e4fe1dde6f3077d125a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/736.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/736.wav new file mode 100644 index 0000000000000000000000000000000000000000..12c9d4c0463bd59f2d4e004129d678dafae66ef3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/736.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d53fc4503d282a225bd02c35be02717856b39058d1eec3d42b8b6a875230c96 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/738.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/738.wav new file mode 100644 index 0000000000000000000000000000000000000000..56c888f17123946a50402e1e749a46d951a0972a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/738.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e4c0296bcf36041496ed301ef450a9eef4bff0c03bab54f524da962ad380d71 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/739.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/739.wav new file mode 100644 index 0000000000000000000000000000000000000000..3bbf48d44b57edbf27e687957ac7949faa206870 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/739.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccdd0d5cdc3d76618650528acfe0601a671bc8aec31535f5c964772bf9e20193 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/740.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/740.wav new file mode 100644 index 0000000000000000000000000000000000000000..1001126de326129e901c20b4e186c62510650bd2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/740.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:560862ae4f8730a66bd4d7e9b7597303aab21f7a98d0f77fe7d72d4805193a5a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/742.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/742.wav new file mode 100644 index 0000000000000000000000000000000000000000..6751b32c7dca53c1b60ffd570ea97bf090afb154 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/742.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e6522a46a642f2c5dad5e3bc96c0582a8b86a45f665c755018c39f4ace52d0b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/744.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/744.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee87551b79c2e1c46e72ea94fe216ef4d0dc81ce --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/744.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac11799645831f78b8a06f740e70725f5e66088f60db810174f3e7ec78528a8d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/745.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/745.wav new file mode 100644 index 0000000000000000000000000000000000000000..6aca8726e6c3e63fb4839c04ed20b5c888ccedcd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/745.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b139505a7fc7d48af421590635d75ed2da15ab4a1fbdf8866c89aa21ab92b62 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/746.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/746.wav new file mode 100644 index 0000000000000000000000000000000000000000..6ad89caa59198390c2e7e6e8a1a504937aceb92a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/746.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a4aa87354a493894124b2901437d87638633d6005c2a5c32c969bd01382c4a5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/749.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/749.wav new file mode 100644 index 0000000000000000000000000000000000000000..032dc39e309a5f58f758a0ac0f1d55bc86473010 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/749.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8df381e99cb2522143747a24e29b17e1ae8a56b2e3a8f3bf14b86c6952324abe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/750.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/750.wav new file mode 100644 index 0000000000000000000000000000000000000000..b5e4e1ae1f85f40ab09c89965c30d95f2015c062 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/750.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc73f3b5e38ae589fad16c7173328a2e25870d71b4b7963c82066a3b18bf8061 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/751.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/751.wav new file mode 100644 index 0000000000000000000000000000000000000000..65d5f24fb87700dc18e3dcb54cfd910efdbdb6f7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/751.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e8eb85b6f1ed54343167a6ba30676626135a2d06abe7c015d30585e44d21474 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/752.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/752.wav new file mode 100644 index 0000000000000000000000000000000000000000..b4729ed349e60684ea6780884c6ff4c0c7114a14 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/752.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ba3f138e4587ca369645b282bfd8d7fec701b92c0ed584eb25a3f054a4f502c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/753.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/753.wav new file mode 100644 index 0000000000000000000000000000000000000000..873e10fe0049c4c53c35bf5ba60248c90a84fb53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/753.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b82ca31a1d58a21f518ffd850e0f46ca51ae0e210f09077fa869480a8dafd2c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/754.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/754.wav new file mode 100644 index 0000000000000000000000000000000000000000..fd23707d743c2fe4ea52a8e2b91f0e553b156ec3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/754.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ccfd3e30ee3c9fbbae51d37f53e75dbcfabf4826678f66f6e862352e2d84b74 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/755.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/755.wav new file mode 100644 index 0000000000000000000000000000000000000000..ec722374f72a2939a0904f03fdc44b27154a1919 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/755.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ced3139eb69d33c99243a69b24d50c86af36b93021c35504abf4ab4c9e6b08f3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/756.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/756.wav new file mode 100644 index 0000000000000000000000000000000000000000..6673fa8e8d61bc0dd35133a5c90ed9ab9270161d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/756.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af4048901351d9f41545f1f8d10a874d9f602a39e5e5d1aea6e24339a6706f36 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/757.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/757.wav new file mode 100644 index 0000000000000000000000000000000000000000..0778e1963f51294cde37cba82e0a0c1265d686b1 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/757.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f93168abe40bbe902bd19e9c8c3c7efa0e271913883492699b15967592b5efa0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/758.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/758.wav new file mode 100644 index 0000000000000000000000000000000000000000..8cc571e34a7b0f411692db15c1863bdf180b0efd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/758.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e793a63653a6e875b518be6b303e64962bc08472f08a7df921aec70cb90571aa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/759.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/759.wav new file mode 100644 index 0000000000000000000000000000000000000000..3cca342e4cc2ff87d312e2b126243468ae2aaec5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/759.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71e21b29c77cd39080c93a70af6b53ec56cb9e2d4ca8e8c35078fe9d26cae0f8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/76.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/76.wav new file mode 100644 index 0000000000000000000000000000000000000000..cfce12f4f7814009410196cb213641dde1c2f15f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/76.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:127f7bc0aa18cd4017db7d83ac1bb2d40fcbafc02564ed218048df0d9a115884 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/760.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/760.wav new file mode 100644 index 0000000000000000000000000000000000000000..880fb974890826ae643886d4b3662f50d0c2aeeb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/760.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:203d8104ff9cff42ac941e613216d8ed0a6ba36889488906ff4a23b5db73fff9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/764.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/764.wav new file mode 100644 index 0000000000000000000000000000000000000000..26f8e4f84ddfa3369b7b3eea9c1e771f37dc9de6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/764.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a15c7fbf38bdec8b4547748e7df49f41df777a56de99c9bba6c0022039f0b6b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/766.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/766.wav new file mode 100644 index 0000000000000000000000000000000000000000..82b01a73bba9889a04808f53d0f9da3d96d5a772 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/766.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f050c080147aa5d0a4e93cf63ce91fdc5d73f5969161f4b29e80db27ac38f7d1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/771.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/771.wav new file mode 100644 index 0000000000000000000000000000000000000000..d3defc66bec0f64b8ee5832848ab1fe27feb13e6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/771.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c24d3282efff1a92edce743779c5b4681f08831ca3835eca16df432a83e12806 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/773.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/773.wav new file mode 100644 index 0000000000000000000000000000000000000000..dba055fa1c749e330fe5936c2a6855771624d4f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/773.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56129b5e5f2fa883b2510f384fb73032f1eba01964b7e09afcd7f95c4ba47f98 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/776.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/776.wav new file mode 100644 index 0000000000000000000000000000000000000000..592873ec3e414f78169987abbb0dca44c74abe97 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/776.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f4748754b2f4b5b22b2960555ab2792e2d75e058fec3892c7dc0d3dfd019c95 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/778.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/778.wav new file mode 100644 index 0000000000000000000000000000000000000000..06f9374e2f40594274161e00171462b7f7c70c42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/778.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6911f26eb62b827dadd265f66dd759c7c7a6be1c4eaceec07379465bf2ad6d9b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/779.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/779.wav new file mode 100644 index 0000000000000000000000000000000000000000..4906941976ded7d6ae939e05d18a03b03d34b08d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/779.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a67fe35966d49e4f71556f8d90e2bb7ac2df59c44a185bd18587df3bebe0912c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/78.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/78.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b779133507be58922b3545a55543e81a1d1abfe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/78.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd6491c062d328a261b608a5ca599357830f7c8c6b9c87360c1fbe62c6719c57 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/781.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/781.wav new file mode 100644 index 0000000000000000000000000000000000000000..b704acc0d98a67b28c767d854c6f91af80013b4c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/781.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dd7853d436cd4a09a7f021377d666c9ba6467a420c9ff45f552c9b8bdf508c4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/782.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/782.wav new file mode 100644 index 0000000000000000000000000000000000000000..f15d4f72d4d9d23e3a38e76b9236d193dcdcc0ee --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/782.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95bddeffdc36ed0553348e2971daf81b12669d79640033fd2b3eedea8faf128a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/784.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/784.wav new file mode 100644 index 0000000000000000000000000000000000000000..178c7ade237c1c878c6a2144a0d8ac822d0ac7e4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/784.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eab4af9f41076ed7ed8cd2f2a83490e1cd4291abcaeddf333bca63f39bd8bc07 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/785.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/785.wav new file mode 100644 index 0000000000000000000000000000000000000000..22c97e6a9c183efcd111e7ec0fceec7b8cc52ddf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/785.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5043afbf209584215ef8256378e72bd0fdae6991631424b61042eb16c471c63f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/788.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/788.wav new file mode 100644 index 0000000000000000000000000000000000000000..292648ab7e8c89cc9d0d22268cb42191d8a0401c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/788.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cff4dc93c0f9cf14fae7595f38d5d8314b6070161b09416d902862657214b6e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/789.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/789.wav new file mode 100644 index 0000000000000000000000000000000000000000..a265e9abdf79d8cc9835be611b770a6122c3065e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/789.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1dad0ef7b2ad26bd9e4bb496eefac952f402115863fb5841e3d6bf2d4b0ee586 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/790.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/790.wav new file mode 100644 index 0000000000000000000000000000000000000000..bc426f31297cf1461b5188de93310ff1d0066846 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/790.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cc798c0d3e5b73d7be7e1a6901d03f4be78d990ea7235ed061ceae889dd6ee6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/793.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/793.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b25e5c20005b51212a9d634fa76c65a6f6de7c4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/793.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:536c0e3e9beccc1dd4bc0dfe855311960344af0b2444cf970097dbf54135539d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/794.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/794.wav new file mode 100644 index 0000000000000000000000000000000000000000..542855bf3cf4186483d776e253d3a77982ebfcfc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/794.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1341ba0dc20c6d965248186f0f20279c215857bf7965452132dc2eb688ba24c5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/796.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/796.wav new file mode 100644 index 0000000000000000000000000000000000000000..de2e002b1a9da94dd8465b6c648c9cec93a22efe --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/796.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d80b2cdbe52a2fa85aad2a211741f44aae74b2bcbc52f651fd57dbc9956710d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/8.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/8.wav new file mode 100644 index 0000000000000000000000000000000000000000..331d895619d74f8cc17cb42ce05d26ebcd6f9a5f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/8.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d0f31d95c8db2512fe979eba5cd11c5276c98507ca55d0551ebf2107cb024aa +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/800.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/800.wav new file mode 100644 index 0000000000000000000000000000000000000000..073ead6a9c93309bdc6e823879c73a5acda91c76 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/800.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f5165b3d5419935dc4813f9d7f9577ca1cea1d92cf20fcc3ed84c7649f231c7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/802.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/802.wav new file mode 100644 index 0000000000000000000000000000000000000000..2e0b6cc3410fdc7338db0e0444b03e0e5a83a369 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/802.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18b127d55b64dca175b36817626b11786f96db3435748cd80f082b286ce113eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/804.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/804.wav new file mode 100644 index 0000000000000000000000000000000000000000..e94d73a52d643849c8de9aef461710d17afc141d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/804.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:186162ad1b42edfb19a305283966919a8429442f8f555e9a817489d6be96ee24 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/805.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/805.wav new file mode 100644 index 0000000000000000000000000000000000000000..06b66c1fcb5a34a9123a9d449722b415e5493f68 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/805.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4116c97354fd7fd5d8179f3f85e66b534f5139e6c3fc417c58f865360ee60a3a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/807.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/807.wav new file mode 100644 index 0000000000000000000000000000000000000000..915e8f1170b43a8495c4287c0f9b87f400c21da4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/807.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:413bd491b198c49fbcb0f40e5794feb8bbd55fdd6f2fd698fe7572b8ce320638 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/81.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/81.wav new file mode 100644 index 0000000000000000000000000000000000000000..c9249a2ad6d396cc26a2e1f4ccf127d446119805 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/81.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73e43c7b4dff78d486917f992f0c3b93907508a3710fe15295a71f8bf38ae2ad +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/810.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/810.wav new file mode 100644 index 0000000000000000000000000000000000000000..37d04cf5a740770ddeb109fd75bd9560c99c7c68 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/810.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0eb29113efb7d97852f4be26973dae31c55ec68b5d8414c74414299d5d881e81 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/811.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/811.wav new file mode 100644 index 0000000000000000000000000000000000000000..7bb00a1506003f2618e59d0759ee8e599a866741 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/811.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fe969072f05c7252b6c0d611087eb6bec679dfcdbf3a8db979042e6df6c37e8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/813.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/813.wav new file mode 100644 index 0000000000000000000000000000000000000000..a4fd08cd32a39e00d78af3dd083291e71b3f9e00 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/813.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd05529c95cbe14c612a5ebe215b20a48a02262823b96ed184b9158adab79151 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/814.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/814.wav new file mode 100644 index 0000000000000000000000000000000000000000..4c37a301e12bdf9163a50d2afb6f8c2453f14fac --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/814.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64eca4973fa97d9f7b363791cc16929efe8ababfbcfbc303906378066ebb1ab7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/816.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/816.wav new file mode 100644 index 0000000000000000000000000000000000000000..70fa1f9aa5811b04ff66e4efafe0f77fb989e42c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/816.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e072ce12cc3cfcb16c8eb31701fe29e20035f919777c22ec23ba7a04b66c3470 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/817.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/817.wav new file mode 100644 index 0000000000000000000000000000000000000000..071fe1a7d941a0c572796bae970cda35ac3713bf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/817.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f6cde93f531b34c5d5e885456340b7463c101eaed89263f46bdd644570d5977 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/819.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/819.wav new file mode 100644 index 0000000000000000000000000000000000000000..77e080498df4a94d7a8979b2a6f4bf22e00b6c72 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/819.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b970d899a2ae4ccffb028e8ca0092df1b891b8c7bb3e0cd0678dc4281a20476 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/820.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/820.wav new file mode 100644 index 0000000000000000000000000000000000000000..40a24b624f03b7dfda54b8954c608c2f0b10debb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/820.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f258d15b5145d570c5540c32312dc950c19c6138567f67d9387ae65f9d857052 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/821.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/821.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1c95d97d8cba2708cd80d61b75df1560307ada3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/821.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a07c58c7da601c45a9a99facbfaa0ca5c4eb5eb59aca23db5fb3e4915d2ee97 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/825.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/825.wav new file mode 100644 index 0000000000000000000000000000000000000000..005be80991297c241ec7541456aaf3f191b4b6a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/825.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e30daa531bf75284d32c72f167f70cd17daca355ba823ceed47f61dd34e0d634 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/827.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/827.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6acf2a3436d80566f58a1c54c31d937656b0a7d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/827.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b7ede5a9d25cfc8b1b6922fd1e32f2511220ced3a85b5bcdf2aa625eaf7885b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/83.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/83.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcea2c8b8a209f6a9ac6296207e8f8e4ee6e433e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/83.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c267869b987f47d5b0b5d0d1b2a30fcc7a6a608c41a32087f334ec04417c9f8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/830.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/830.wav new file mode 100644 index 0000000000000000000000000000000000000000..9a8518cccdeaa283dbfd4a975299c9e04b52bd3c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/830.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdacf4ed78fd5e825527377e1571d96566c3d71a4e37e086ddfcf21408ff692c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/831.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/831.wav new file mode 100644 index 0000000000000000000000000000000000000000..c920c768b965ae7afec93f6935651da25fab3a5a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/831.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad108fa7eca362799ea23c4af58655cff3c983f20e3a488e6711da02a7fe498a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/832.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/832.wav new file mode 100644 index 0000000000000000000000000000000000000000..834575553f29fa57ae7d8d300ddd54951461f849 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/832.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e51914e10ca4a36e1e987a42665f2fae27f24f4bbef2a1005bc04dc2719c4fc3 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/833.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/833.wav new file mode 100644 index 0000000000000000000000000000000000000000..11a44912a251a2cca00306421aaa5e9738cd3944 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/833.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384b8f0fbf4fa0a2741f9e2844281270246e390d3fe9cda145c478684266189d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/834.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/834.wav new file mode 100644 index 0000000000000000000000000000000000000000..ffcc1288f538629cee0e1238adb73e8848372068 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/834.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a6aff2e6a319ffbcd6a6d01014a8b9541982d1d133b7f4415236c6fd86fad40 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/836.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/836.wav new file mode 100644 index 0000000000000000000000000000000000000000..65ca3f0b5cb52def2d8d8b744ac8a453a9570f37 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/836.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e2bf00444dea0e4ae51a3137d1ff276a2f2a2d273af69d54651cc77df74bf96 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/837.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/837.wav new file mode 100644 index 0000000000000000000000000000000000000000..0ae427c59ce34f01077a735560eb851fe17fee88 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/837.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e49225d79bb0182e07fc27b476a986ec3d1654dea4fe9b432fec6da119bd10f1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/838.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/838.wav new file mode 100644 index 0000000000000000000000000000000000000000..e13aa1dfbbb4546490f391f785deae8f8f4510e3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/838.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8fb8d4a93307949fdda9be89c9a0e983af3707cd48f2bef028fa955d5639bc6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/840.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/840.wav new file mode 100644 index 0000000000000000000000000000000000000000..281bf09cc1b9910848d7dfeefae56e25db511635 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/840.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:237fd58bc8e0ddda8b053c9f83950acc094d5326579ce4e77f030869227fb56a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/841.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/841.wav new file mode 100644 index 0000000000000000000000000000000000000000..7ccbfce133a8f8e057680ea553d8847bc601b1de --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/841.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ffed92290291a4f00caea7c9e0bfcd6d2e04410edc75e1c348be58d01f46ade +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/843.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/843.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0178d4a814729160469e1cc981009673c091560 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/843.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:555817623d938068910e69375f04952c56c945e123e2925be3b099f09b17f0eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/845.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/845.wav new file mode 100644 index 0000000000000000000000000000000000000000..c7d0e8eae04481b2087da1c9f013119406baa66d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/845.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:20bfa48cfe6562f9dd03fb77e54602ff4296f3514cb60e89e59864fb336f08ad +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/847.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/847.wav new file mode 100644 index 0000000000000000000000000000000000000000..35b230cc891ef788834ee61125547740aa735808 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/847.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a94f0f0b3b933e407588dfade862047a1d821fde143a49317ceb3175016f39ad +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/848.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/848.wav new file mode 100644 index 0000000000000000000000000000000000000000..9690a866c559674cd138c701ec0e4088a85ebc05 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/848.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78f8912d79cc79e85f160f8444de425a5ab9541488994d4d85d6cbb7c95ac610 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/849.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/849.wav new file mode 100644 index 0000000000000000000000000000000000000000..13407afda3b517fb9c9827a6fde1bc3adb151957 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/849.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bc8c8f91c475ff3960b748426850746a60026513ed54f543254c059ea5f7de0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/850.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/850.wav new file mode 100644 index 0000000000000000000000000000000000000000..98f4bd17754416d6e8eeefbbb0c51643e7c4d7a0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/850.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:612c41dc47782a0a4087f6d132d4964f0158167e82717f64d4ad6f3ecc60c4af +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/852.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/852.wav new file mode 100644 index 0000000000000000000000000000000000000000..4570fcc86a99a18f5dde8379b3eeeb1c9d9976bc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/852.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08e95adc77be3062ebdb74b439e79a80bd860e34368588737cdc359b47b81314 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/854.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/854.wav new file mode 100644 index 0000000000000000000000000000000000000000..faa9f67d3650eb43fe212c0d4592432d03e2ba61 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/854.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce179ba72ee12886e6ba434a75a3a0f1f914b9a0d6174adc71a29398eb65124f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/855.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/855.wav new file mode 100644 index 0000000000000000000000000000000000000000..05ef44e685cd4768a8b4c6c9c11ae8a9104357f2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/855.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9deb6fe2a07a996413d5b39b4e884bb0ac321b83d8798f04d3ada98dd4938b1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/856.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/856.wav new file mode 100644 index 0000000000000000000000000000000000000000..048db0616eb8e2c6772c651d9df6a5bcaa10a59f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/856.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b34dfec61a229b627403fe1d240d62401ada2638468bb538fed9ca7a667ba6b7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/857.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/857.wav new file mode 100644 index 0000000000000000000000000000000000000000..e36f9e96b7b4b2546ba90a5db014034dacb48d54 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/857.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80fde0e42188bbbf7989a950e90eacfacf414e376367ef59bfd0fa1143b4a41f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/860.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/860.wav new file mode 100644 index 0000000000000000000000000000000000000000..4062207fabc7f020bd7a008b896056b640bf2bc7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/860.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e073e9d3adc7035225db40fcbc03785f593d3320e29c6556ffc590684ebdbf54 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/862.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/862.wav new file mode 100644 index 0000000000000000000000000000000000000000..da7daaa17c77dcd10c0728ea28988d1bf8725eb5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/862.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffb6552b48dac0386f27bd1df408b58e369f8f3598e8bac0661fb59ba1c1530a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/863.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/863.wav new file mode 100644 index 0000000000000000000000000000000000000000..d7876258ed482a6b41e6fcee53355095c75af9dc --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/863.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b73676083000f33f0873c09504cdd1f71f529b63a42db3807eb176893ecb5855 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/865.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/865.wav new file mode 100644 index 0000000000000000000000000000000000000000..4934c0f222796aa0718431c6a7a877dd9b9c5522 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/865.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03f89ece6908d76a4dcf1dd62f44e032e1ed765f5ad655c7df23c16e5217da94 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/866.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/866.wav new file mode 100644 index 0000000000000000000000000000000000000000..2f17d6d5cf1317774b0438d4fab1b73071d5b1f4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/866.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:401f7166da3a1f970bc4079e19c4611a06e0688d06466abc34bcfc18f80e7ab1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/867.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/867.wav new file mode 100644 index 0000000000000000000000000000000000000000..ea0e0b764c2e9a2e72a87ff66bd7b401b24d8803 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/867.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:254c3fc7abd82ce07e4502add6c62178b5572a655efc2ab7dd26b2265b9da01b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/868.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/868.wav new file mode 100644 index 0000000000000000000000000000000000000000..4b1ba803996688aa978568725cebd39dba4db54a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/868.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8dedd46ec78cce2c42c6db675f93d17138e53bb277c9870c90026b99bc68eb7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/869.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/869.wav new file mode 100644 index 0000000000000000000000000000000000000000..00d862beb82cc022f86a78784c38230f3e1fc562 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/869.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aad4a2e6194fb9b71461938eb5dc376f2cc9df049c2bf3bd07705140e9bcfa16 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/87.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/87.wav new file mode 100644 index 0000000000000000000000000000000000000000..8f54dfe4672cfb5ac22cfb00454a0fdf8846c36e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/87.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09241de3b1580cc773be08ba21ef9fceb851a273680cd948107ef21d0a377c2d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/870.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/870.wav new file mode 100644 index 0000000000000000000000000000000000000000..918bd7c7746e96b476ad57c7ade157fa09ba9901 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/870.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ccd0475945fffdaf67ef30de62831741f165b00e79387bddea30f480eb31db6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/872.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/872.wav new file mode 100644 index 0000000000000000000000000000000000000000..f342af359f3c11b11f3f967363070800439c2236 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/872.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf6a16d60ec4c70d5b390313d21e63c367b7e6e85316df9eb7315d74999f65b0 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/873.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/873.wav new file mode 100644 index 0000000000000000000000000000000000000000..db8a1a3bdeb4b3eea17c87528d26f62ee7839280 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/873.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72576b468e5352f665763c6a254935e3440dbd059b749ad4d65b21377fa5a361 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/874.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/874.wav new file mode 100644 index 0000000000000000000000000000000000000000..bdd18f4e5defc788cc0c2120627e36b922638ac3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/874.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce83c675cfb7df56e2ce9b7b00c4819d07fb6933b9fc9f75c29c2fe515bba9a2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/878.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/878.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1550379967be067b02f5b64c5a1dadd1996bfdd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/878.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ea737b21470a04aabe348a0b9d4230afd2b19cc7e6d3ad7307b0f99a5bfdd42 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/879.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/879.wav new file mode 100644 index 0000000000000000000000000000000000000000..a21db832fd622fb3618e2e78823332ba0ac8f746 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/879.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0874807d6b779a7e289905b5e43804e48164a564b40660d491057ef06874e44 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/881.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/881.wav new file mode 100644 index 0000000000000000000000000000000000000000..97a24aaa9f5b8f06763a928103215087e4c4abf5 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/881.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db6eee1cee9c1859780988baeac6965bc0f0176f6e1c9b1a3ce068b3e225ffba +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/882.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/882.wav new file mode 100644 index 0000000000000000000000000000000000000000..016fef8e6b9610b2c9739a1fd2a702e1c807bbc3 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/882.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a10cdba01ded68a641f756d48db701af94a61c65979ed4e40708f97f7abbacc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/886.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/886.wav new file mode 100644 index 0000000000000000000000000000000000000000..8c5c1e855f6e115d458efcc9ca44bb7db48e7839 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/886.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f778ab0b2a4c4a9b8a581ff8f1aa5025ae34fc6f539fb5243ae0781f3f1849bb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/888.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/888.wav new file mode 100644 index 0000000000000000000000000000000000000000..6b892cacaee358002613feaa4b7c8874142a1597 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/888.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3bc4d211562a348f6af348f0fde2339a0ad957be072522251c1dd8643bc039a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/889.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/889.wav new file mode 100644 index 0000000000000000000000000000000000000000..ef3b608be4b0c58725e8b4636da79adeb1f8f4dd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/889.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b18f60aa29ba1cfcf804f8f7ffc6d448462effdba3517d0fc032da3b384c4e4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/89.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/89.wav new file mode 100644 index 0000000000000000000000000000000000000000..f62cdc9c7d66ffa5519c555e7f919166995eff49 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/89.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ebb6486379e35131865898fac4d7aec0e0d8ead6a7a46dca17f27e80e8fde8a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/892.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/892.wav new file mode 100644 index 0000000000000000000000000000000000000000..68c0256eaff9f604234bd85ac3d6f125646d9aaa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/892.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe6de9435fb4a73091ffa89471446fd21679fa0f50c8f5d60e31473a83c6693d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/893.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/893.wav new file mode 100644 index 0000000000000000000000000000000000000000..5bb1bb581e5807b6777cbe167b31a2d98de3241e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/893.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed5619fcdf54118f94d4b5eec5036316369b01a66b8691da79228321f2c36006 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/894.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/894.wav new file mode 100644 index 0000000000000000000000000000000000000000..a0509eb6f6e04a677e3ad986494711243cbec6d0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/894.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e090f439845b34436ca2fec365e9caa5e50873119fec55b92594d01de4b39b11 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/895.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/895.wav new file mode 100644 index 0000000000000000000000000000000000000000..e0138bebbd66dc03671ef62f846f282fd764a512 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/895.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79d4cadd495197231ed0f7a02b9141f23d295833d7be663e3e42062b9d152c20 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/896.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/896.wav new file mode 100644 index 0000000000000000000000000000000000000000..a6d6cd7265d986daca0d61ab3469917588c607d7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/896.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bce97431689c4b0540bdce2341b48aaca94e759c6948b587e664749c58139508 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/898.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/898.wav new file mode 100644 index 0000000000000000000000000000000000000000..fdd50edf087e3865d38c959a60e12b0667d1e096 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/898.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cd7396fb7b83d4536914aa3709e8943cd3d6a63b08b791e4b8b8b9165af93fc +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/899.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/899.wav new file mode 100644 index 0000000000000000000000000000000000000000..ee4ac2c5c4f083b9af74fac943f8867b0a38a6c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/899.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9ac9dfdf55eeaad1e1ec6af190d6481cb606ce2e7eb31c2db4a141b7b7c89ae +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/9.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/9.wav new file mode 100644 index 0000000000000000000000000000000000000000..adc6469ef261770bc90c39035fcca707f9576efd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/9.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15976d28bef04c9f2365b9f60bda81d1e98e49e133cb977b22889d2c5164e4eb +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/900.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/900.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcbf0c6e053c16ea1e53fd2d21f4bd043217e49d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/900.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed9dee5c8f3cdc4330ffed11d988f1b55d6d4808ca227f7a8a28e0f15b3ee720 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/902.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/902.wav new file mode 100644 index 0000000000000000000000000000000000000000..8b35d2f1973d6712274589dcbc2604ad3dbea673 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/902.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7abd6399cd39955f6ceba577d9156358e5240ef9322677852b40bfbe4a58921 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/903.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/903.wav new file mode 100644 index 0000000000000000000000000000000000000000..c08afd3e7c37387ce885c2ef96a06cdea1fd87cf --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/903.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dea9dc8a6bc219a1bf116f4c8c11138bb7a69b4bb55ca5eb6425a1eb3d4409c9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/904.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/904.wav new file mode 100644 index 0000000000000000000000000000000000000000..9ab10a229a0e381fa7128566e6649a5d370e43b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/904.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a67576d66fc5c61298843b3101112868afd1a8874e7a3c0a88d36dbd0e777e6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/905.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/905.wav new file mode 100644 index 0000000000000000000000000000000000000000..8269bd7955f6fa776b5c71e20b561ba9c567dd0c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/905.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35d67fb10ff593bf7612400a46ec9bc66d68f2ddaa50448b4afa2e2adcfb65d6 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/906.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/906.wav new file mode 100644 index 0000000000000000000000000000000000000000..89e21c3e9137f51abf02e21617a0722b6f1e564a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/906.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86b20a42e66033f5e111f50cd24f88085cbf6beb25f65995335fd152595a2ae4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/908.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/908.wav new file mode 100644 index 0000000000000000000000000000000000000000..c82b9b2084975c3f41de40521cb858a441825d85 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/908.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39d14db2b82edd8846a3c23387d9937d29d1c4efc274737f42eca157b2154012 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/91.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/91.wav new file mode 100644 index 0000000000000000000000000000000000000000..e9a1b7c0244db8355a8e77439753124903829033 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/91.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50247ac5f156a3fa08cc3596e8e75196ce2d29031a4f736ee196c1d2da612090 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/910.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/910.wav new file mode 100644 index 0000000000000000000000000000000000000000..777f20019760177a34620d2d2e8fb4761452fce8 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/910.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33eb08a45fa264eb3c30caec142bbb2f3bfaa2465d7a5c424e3875eac9737839 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/911.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/911.wav new file mode 100644 index 0000000000000000000000000000000000000000..fafaff708e6e6c9b3503ff3dbeb12516aca6afec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/911.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:084c936270834ad607325519abb5dadaf9fc841f47254a66a5cdd5126c31a536 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/912.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/912.wav new file mode 100644 index 0000000000000000000000000000000000000000..dcdf6f951a7394487450c57d841144bce9e1f6b2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/912.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b6b47a8dc5cbb60163be66969eca73f391fad056ab5cb46c86b9cf68fb7313fe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/913.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/913.wav new file mode 100644 index 0000000000000000000000000000000000000000..3ec745b46ba61062b290afda0ce29b3dadaaf484 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/913.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8344456fb059abe816f31b615925eece176b7a6d0248de86e079b0799883176 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/916.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/916.wav new file mode 100644 index 0000000000000000000000000000000000000000..27eb97f65341e771cbd22a9e4a2ad14580ceffe4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/916.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90d22e3a5d90ec97640c26ac4aa27a3709988a82c928e3974d5da7a4f01ad58e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/917.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/917.wav new file mode 100644 index 0000000000000000000000000000000000000000..f328b73f12de435b149d1b061e735d76495a4a72 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/917.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6a4e51dd06119226798ea8ada67b488e8007d9d4468bfd624d8ad369fac3e04 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/919.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/919.wav new file mode 100644 index 0000000000000000000000000000000000000000..e239e86d153cffd7c65b3e055fcf28db644db593 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/919.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec998d57b520fcd398885ebaafb592a1ac3f111a168c1110bc740ea57cd389c1 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/92.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/92.wav new file mode 100644 index 0000000000000000000000000000000000000000..1e981f3205dda49fd4325961d8f9e68ea05fff78 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/92.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97ab779db8e4403a4c538501f34738f43497eea55bf895f46d72beaca90b594a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/921.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/921.wav new file mode 100644 index 0000000000000000000000000000000000000000..cd92a99ac3a498ba497ac5f692b2f5fd09d7487b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/921.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd8b6ff913ad98a3cce6e944d208033171e150ad372562a17df7cb6d11a39f1e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/922.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/922.wav new file mode 100644 index 0000000000000000000000000000000000000000..983d884ec617fa3c19a65568025437e885f1ac94 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/922.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43f6ac392c30fc78f121565d3f8d5549e9fb3d1cd46edba8d92b0c80c624754f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/926.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/926.wav new file mode 100644 index 0000000000000000000000000000000000000000..f76356496c8d7fa47ae3d77eff63c7d117669d54 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/926.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8fbf8e1ce5ca16e825ffc4c94d81a402c2e4ba6a5a6f406c5ed336dddf7be322 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/927.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/927.wav new file mode 100644 index 0000000000000000000000000000000000000000..9e0643aab3e895f1a07010dc1a36cccc39fe6566 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/927.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba2688f9b11e7609451490780c896d77f7299c7616221150ac489f8b1bdfc01b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/928.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/928.wav new file mode 100644 index 0000000000000000000000000000000000000000..648dee47433e4356c661dcd9329ef88123e2cb5e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/928.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:871830082c305b2241fcbd26ff89be94af256df33906a3425d2016771d958b23 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/929.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/929.wav new file mode 100644 index 0000000000000000000000000000000000000000..dbf2b32d1b36aa4e39842dcdf225324f5427e776 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/929.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0de8aabda37dcebb277318494592fcef50b071e9f1d6ace540d045d4e51682ec +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/930.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/930.wav new file mode 100644 index 0000000000000000000000000000000000000000..61b9a717e42c43a45aa094de6c0394a087798aed --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/930.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64268274e5d4044b1116f4cd30201976fa87001dc6c889d5c0b52f3cb8334e73 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/932.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/932.wav new file mode 100644 index 0000000000000000000000000000000000000000..542acc87029767a180bef08b6d1fa0fef7047598 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/932.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a3e334938de43b4dba996277838427f47ea4f41d8546ae19726e79e287a92e5 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/936.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/936.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebd0acc4deccef167d15bb62de51e1b8b24472af --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/936.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8089d9c3ce9002d2acd521f29f093ea4e341b99fe9cb9fcfa5e453dcd998c0b2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/939.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/939.wav new file mode 100644 index 0000000000000000000000000000000000000000..3c1b65ef09762fa9fde84c163e3affaf3012b1a9 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/939.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3e7bdecc7847f7ed101f9149b7a06c1a48565110a7a44e7f3a78c1d7fbebb84 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/94.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/94.wav new file mode 100644 index 0000000000000000000000000000000000000000..6794c3269cda2a52f71c1b1eb759b96f889d9ce7 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/94.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6af4d307424b807a9c093b9cde38e962d53df93373ab8c31fa0dcb6eaae7ae3e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/940.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/940.wav new file mode 100644 index 0000000000000000000000000000000000000000..1b74b4afa6c0c8279f6de8ab2cb33fc1eefc775e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/940.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:47db0ed5356d9e431a0c88efa805f51051fdc0f300d1b287c1751b4e96dcc8fe +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/942.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/942.wav new file mode 100644 index 0000000000000000000000000000000000000000..219424e99b28f2c40e6b33a53a17db50d07e539c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/942.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:537759806d8a80a96b30584927e7d82bc2f093d18b9717a1c952c54675d6d18b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/943.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/943.wav new file mode 100644 index 0000000000000000000000000000000000000000..01ba9cbba2543f179d16ec0579b1e3661ff49ddd --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/943.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ffc4a839548f1169a1ee29237f9e38cc3d40ccf89951bc427c8626cb4aa0d3c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/944.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/944.wav new file mode 100644 index 0000000000000000000000000000000000000000..3a05643d1ffbea318f0b2ae46ce049d5078d9c06 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/944.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:193f3de53ece205cedf79bd84076d78516c9aa6d1e705bddeb4f4c3801dc56c4 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/949.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/949.wav new file mode 100644 index 0000000000000000000000000000000000000000..47f9d5d03e2c90f5c97b4eda603c264151426fec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/949.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70770e6549d7d12c6c92ac5191552536986e389b306f3059ecbc3b802fbbddf7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/95.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/95.wav new file mode 100644 index 0000000000000000000000000000000000000000..4262ed365fa335369b4500c4c312304c083c51fb --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/95.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46dacbe5f934a3a59654c38bc54b42dd1a87e6b6b933281cbefa8780b60ec468 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/950.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/950.wav new file mode 100644 index 0000000000000000000000000000000000000000..b0e5f8e5e1f9887f8c92c42828b637a14c064e3b --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/950.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5c0a35d379f5a20c169a60a0e3c4971b8ccd70e96073ce012bff3a8dd32db0a +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/953.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/953.wav new file mode 100644 index 0000000000000000000000000000000000000000..a955a9de3f7f1cf38e353e9df781942b18ce720d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/953.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03b4b9177f061b63e22f318e7c681f971074948e8bcb36527da023b40bb49003 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/955.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/955.wav new file mode 100644 index 0000000000000000000000000000000000000000..32254c45146f6d6522e884e7312a4d23ece594c2 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/955.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d4152dd34b0bf590759b6e1824f2c221dd70759fa176ed98e2e428a1be1024b +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/957.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/957.wav new file mode 100644 index 0000000000000000000000000000000000000000..4ac393308ea75d2ba8e86621231d7b537e472c17 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/957.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fffa0ca266c482ee95a696624977ef6ddf084672166d2ad1e20b495918a4436 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/959.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/959.wav new file mode 100644 index 0000000000000000000000000000000000000000..230dc64504e255280d4e0b9b951033fe94413c42 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/959.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf0351855d03f311dd47a4a2cbdd85fad4bda8a67de88269e6567f9ad83d448e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/961.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/961.wav new file mode 100644 index 0000000000000000000000000000000000000000..2cf58c60348f5a287809cd3c35c07f8e78fb3084 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/961.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4961086b74ed6225f800f9dda9be0ea268247a718a84a24445c565f316e6ce8e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/962.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/962.wav new file mode 100644 index 0000000000000000000000000000000000000000..fcea2b6adff75012129a74332ccbb7b4fdd86c5c --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/962.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9a24f38cbb9774d5652cb4bf188bc83e19520fb7660114b44c3427c115cc3f8 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/965.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/965.wav new file mode 100644 index 0000000000000000000000000000000000000000..be5ffd474e427b80e2fb4b79395714782603499f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/965.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e735db818b2f55cffde9dba8823c8e9e0ee9e080c09884e9fc68bc249067d09 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/969.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/969.wav new file mode 100644 index 0000000000000000000000000000000000000000..28247a261c07f146101f6a846e012c6d7c31e59e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/969.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03cad324e4e4d41daec299bb2961002fb3a9933d1860fafe90fa835ad8daebc9 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/97.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/97.wav new file mode 100644 index 0000000000000000000000000000000000000000..d1111f035b9120e4dcc9915b2be86d34bf6c703e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/97.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9bc163435ad7c61def4c2a1ae0ecb7bc4f612db4cce986b57c0f4e7513d4fe7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/970.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/970.wav new file mode 100644 index 0000000000000000000000000000000000000000..2455584ac7eb647e84461c30a1601c5458bb0001 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/970.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:14d014599c387288066c921808be526a14f8d269852cfe6973a3a7532ae831ea +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/971.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/971.wav new file mode 100644 index 0000000000000000000000000000000000000000..7ca3a406b4af043c4dd5e6a09c1fa95680f2029a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/971.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b25c485d84f095e0eafbd1688eac15bf48761883e9a5f432f8b46fec5b7c8887 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/972.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/972.wav new file mode 100644 index 0000000000000000000000000000000000000000..eee4090c4f9f59158afea01cc06104b603b0eb79 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/972.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09d99ea62a97cc548becbf7c7983f580dee713ad9f0ea648a288d60d2403ae8d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/974.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/974.wav new file mode 100644 index 0000000000000000000000000000000000000000..f88675c0f60cdcefab79231b13b509773ec28f64 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/974.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:68cb7948311d014a0c41dce5ba96654cc64bf00ae6269de35ee0bf4565ab526d +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/975.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/975.wav new file mode 100644 index 0000000000000000000000000000000000000000..5831ca2c8e55b0eb827e8fcf6df89f1289c214c6 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/975.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cf83eec0befee145f68df1d108a6040024226e77b149e396d9be0d54c70b147 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/978.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/978.wav new file mode 100644 index 0000000000000000000000000000000000000000..f8b3369c214af05ba87f9685ddd249cc1db63f53 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/978.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d9296e77524e6de4f325618b1014948ff9221182667004ed951d2e6855a1a41 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/979.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/979.wav new file mode 100644 index 0000000000000000000000000000000000000000..42092b0622629cb94c4525d2bcf99d0422fb1eec --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/979.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4217029c062547f22ec1f2581b2f94c3669dd75130ef9c3fda0c78a1edd5e22c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/98.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/98.wav new file mode 100644 index 0000000000000000000000000000000000000000..fdda9ff9447d94343341eeec0decbb0dbaf7ae5d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/98.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7be1bf1542756743b21990e76d00ef83cde6cdf0449823a4ea995c551a064656 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/980.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/980.wav new file mode 100644 index 0000000000000000000000000000000000000000..999b6af10f7a86cc0f875261a62a32561bf35d5a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/980.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2719026a791b467d66d114472534a06e831ee3413593024eb1b637233b6d82e7 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/982.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/982.wav new file mode 100644 index 0000000000000000000000000000000000000000..6d58922ef492c77a134bb238153dadc0a5f8aed0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/982.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4344191ba349a1b38b4563533fba747a33aa1ac21c80a6ba618cd04d9b05c221 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/984.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/984.wav new file mode 100644 index 0000000000000000000000000000000000000000..ebb5d57daa49d608fe2d399ed71040c0d710f726 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/984.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f2b60c60364c6dbf6ea73ede603e32c53329eb13e99fd08c45a5b1e7de8a53e +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/985.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/985.wav new file mode 100644 index 0000000000000000000000000000000000000000..0854d1763c62886aeda59e65d5babe2f168fb1b0 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/985.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0237b683ea8aadf2e52f0c59b8f5fed8e2c1fcd067cf7a2fd9c6e3f289bd997c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/987.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/987.wav new file mode 100644 index 0000000000000000000000000000000000000000..fae683517c05bab2fc5f2fbed5e5be492bbe5233 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/987.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb2c3220794914434d6676334ce5595f7981b65f4df1581911378e7cb29d203f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/988.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/988.wav new file mode 100644 index 0000000000000000000000000000000000000000..afd338f2cd65436c81aa53afae42edaa8db3d97d --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/988.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:099f9e92bf475e3dcf6c652731484ea417780c200518205bf8f0642f21e0b094 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/989.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/989.wav new file mode 100644 index 0000000000000000000000000000000000000000..691c3f0dee3c2c83a3a3413e433fd0e82a4570a4 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/989.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c309f064d39d965269e2a35a91b4f5af89c5b2e0970af7b7a5f14bec7ca2819f +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/99.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/99.wav new file mode 100644 index 0000000000000000000000000000000000000000..b1aedc77d98ff0880f83563bf6c62a131471b756 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/99.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d05ac71115485d4ac14d484e6821424754135da893b36c668c92cc8c34b07320 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/990.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/990.wav new file mode 100644 index 0000000000000000000000000000000000000000..bf1f4a2128e78cac1aca46947e4387de51da230a --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/990.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e12ddb786e1e29f63e51d62b005bca2c022c43bb023853c46de0b9f45bfc062 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/992.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/992.wav new file mode 100644 index 0000000000000000000000000000000000000000..09f67c04c27e0630fe409a5ace662bbb4e7176fa --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/992.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca80f286cec1999df6cd04110fbece9cace593be96fb9d2701b4fc1ec84f53a2 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/994.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/994.wav new file mode 100644 index 0000000000000000000000000000000000000000..4086243b5e904a94c45210b6c4bdd0e6e41c9c2f --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/994.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18fc0b1786c2245e8f53512261a38f0c2ec47ebb48a1b104a7ce5da1fd76644c +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/995.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/995.wav new file mode 100644 index 0000000000000000000000000000000000000000..b9201a0ca0362c7d28770694b82ad0067e2dfa7e --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/995.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f42e3acaf91dbbdb3b732f625e9cccb274755b7c8944a2bbf784cf036cf56a71 +size 2646050 diff --git a/Foundation/Acoustic_Scene_Classification_TUT2017/997.wav b/Foundation/Acoustic_Scene_Classification_TUT2017/997.wav new file mode 100644 index 0000000000000000000000000000000000000000..c8f786b0012259b64afc57286f78b920864dc525 --- /dev/null +++ b/Foundation/Acoustic_Scene_Classification_TUT2017/997.wav @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bef6036521c5e8da50f2970ca879abfc6d077d56c1fd2b1d5cd62a052874b15 +size 2646050 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-0A1_JR5f34.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-0A1_JR5f34.flac new file mode 100644 index 0000000000000000000000000000000000000000..af7e4763ab3faab6a8bb26c07cd8ae089cda44a7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-0A1_JR5f34.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff76cea0455117ac349fcee386519b1fa641ed2e5a991db3925c9aad0b4155c7 +size 208507 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-1oRfe0FQI8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-1oRfe0FQI8.flac new file mode 100644 index 0000000000000000000000000000000000000000..96b0c95d86de021f9b0bf9be5332233217a9f6cb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-1oRfe0FQI8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1473eae3f927026c2846fe7e09c99e071bc28e51d0ad47ef1fe600051b153544 +size 260185 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-2--Xzix2Vk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-2--Xzix2Vk.flac new file mode 100644 index 0000000000000000000000000000000000000000..7c107859fd426b67b8ce3b4bc9a17b99bb843092 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-2--Xzix2Vk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cd63291003eaff7e694225b30004b52a07de9ed9c29d073b85be9c5a0e075a7 +size 90943 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-55UGnqZsOo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-55UGnqZsOo.flac new file mode 100644 index 0000000000000000000000000000000000000000..01eccb57758f0b380350c16d1e763212cb78f418 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-55UGnqZsOo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:29e4a1b2a5d17dd1d6f087c10169c58ef5ad089608094b5f7f984825c696989d +size 152006 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-5IoPSM59xM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-5IoPSM59xM.flac new file mode 100644 index 0000000000000000000000000000000000000000..1e3125cdf4ddb2dec5f99659056eed6e432fd8e9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-5IoPSM59xM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a54fcaed112bf37f2db50b3b87f5b0a8baabc30b4df42f25dd6dabfbda29867 +size 125184 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-67q0yYR0zk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-67q0yYR0zk.flac new file mode 100644 index 0000000000000000000000000000000000000000..0f72382781591e1d8f3f25c1cba72b4fd39b28a9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-67q0yYR0zk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b53854ee87cb48bb83a8c0189105839916facc7c9d5173b3d3a0cc04b0e467d5 +size 183888 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-68677bCJWM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-68677bCJWM.flac new file mode 100644 index 0000000000000000000000000000000000000000..8ae223f446f5f9fa84b9e7ca05b8af0dbd7cb238 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-68677bCJWM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:929af8291825c91bb5c3d757433085532f0ccba3b0d8a5e86e6e52c896bf41b6 +size 255398 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-68L7q5JlL4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-68L7q5JlL4.flac new file mode 100644 index 0000000000000000000000000000000000000000..b3baa0d9dbd56547ea6a6943cf9fede34584d4c8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-68L7q5JlL4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80d882495be0ee4fad1c902847803517ab64677aef24190ab28751cd5343bfeb +size 278562 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-A3K0_gtK4c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-A3K0_gtK4c.flac new file mode 100644 index 0000000000000000000000000000000000000000..0eac1c0872e7baed63b52bfb76a80d09e0b9963c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-A3K0_gtK4c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2754f722b837aa5b12069284153fc7fef1ca38e279cf3ba4a3fd1d00fbffada +size 292551 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-ATrpUV4axE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-ATrpUV4axE.flac new file mode 100644 index 0000000000000000000000000000000000000000..34706b631b1995dfc1c7967ed020c822ea3333e1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-ATrpUV4axE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9adbced5a6eabed5ef48d0787abba448b9c764ed2fddba8cf3e08824816c65e +size 198112 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-A_Gt60tfC8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-A_Gt60tfC8.flac new file mode 100644 index 0000000000000000000000000000000000000000..2803c465be0ccd6d2c68cd21dd2e01b331d44988 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-A_Gt60tfC8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38ed654a22154a7fb7209e05f6b8fc03fae90a98840732ac83becf6c102b3e89 +size 211350 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-AlpLvWoLKs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-AlpLvWoLKs.flac new file mode 100644 index 0000000000000000000000000000000000000000..918a7a755aeedcf2ed5c68736f3a9e3911535477 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-AlpLvWoLKs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6275dbbee597bf4b885f8610f70ea7b4af3102214dfd13916489e089b67b1655 +size 262845 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-B2i3dzzdso.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-B2i3dzzdso.flac new file mode 100644 index 0000000000000000000000000000000000000000..1d7fd1f06ec5be6f7f79665258c42ef96c91e006 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-B2i3dzzdso.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83e0536314ca2997723a1260f3cd37fb9e6c6207e7667fc087380f02e7cdc496 +size 266165 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-BL3E3sLc78.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-BL3E3sLc78.flac new file mode 100644 index 0000000000000000000000000000000000000000..18315dc08eb4dca2fc25d71e395b29bbc9dbe73e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-BL3E3sLc78.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d213f45a6f2af834fcbc33bafb50c4a712b028639a1a354e28153592bc6dcaa4 +size 272411 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-HVUL5tNzdo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-HVUL5tNzdo.flac new file mode 100644 index 0000000000000000000000000000000000000000..201b6ba37d66a5b0f0c9c43d50d89ebeea39e398 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-HVUL5tNzdo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fe58c530ac6f7a21abd562c32da7b1e251cf76d3641c2febf21d6f536ff047f +size 214565 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-HfyfiDVqAo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-HfyfiDVqAo.flac new file mode 100644 index 0000000000000000000000000000000000000000..7bfbf5b6b70e956e0e6c047cc8ce6179262d16d0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-HfyfiDVqAo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98a1655be857b11e4b9979c467c7dceae57eaf8df5d8873dcf1a84dbec0b344a +size 229434 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-I_Ej65k3Rg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-I_Ej65k3Rg.flac new file mode 100644 index 0000000000000000000000000000000000000000..bd30cc9579057d25b3a03c373b6afef87503bd06 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-I_Ej65k3Rg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b5aa2eb19aab826395505717fbaae570ba821c4baaf77932e26b2267e4adae9 +size 191915 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-JaQwrOeCYs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-JaQwrOeCYs.flac new file mode 100644 index 0000000000000000000000000000000000000000..46257a68d777658cd35af2151e893cebff52fa46 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-JaQwrOeCYs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c0a15eafb16784c017af8873cbfd9b0e4e4b8f823376b7101d1ad2f2e9e63c2f +size 278121 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-KjE40DlSdU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-KjE40DlSdU.flac new file mode 100644 index 0000000000000000000000000000000000000000..36ecd620987cccbceaac5f2f9802aaf6d6ba5d19 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-KjE40DlSdU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c86edec3879a8f857aa8331962b9b04073fd010d34dd1f07bc5bc030b9af3ef +size 239585 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-LKvu9q54f0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-LKvu9q54f0.flac new file mode 100644 index 0000000000000000000000000000000000000000..aa6cc2718f32ab46986269e1173330e6650be93f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-LKvu9q54f0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1b1e4278b3879e6cfa1322370f0afb4921be0deb87d6a007e31e5a6d2eabdd8 +size 218311 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-MZGb2-2DoA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-MZGb2-2DoA.flac new file mode 100644 index 0000000000000000000000000000000000000000..b0f30b054b302ba8af7012a908dff71c477dc860 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-MZGb2-2DoA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1611d33089c793459acbe10c83e4e49b461ac735488ffbeee6995f2110d4c973 +size 276267 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-NsC63dA01g.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-NsC63dA01g.flac new file mode 100644 index 0000000000000000000000000000000000000000..982308304313ea92f178d933b994a9591df21d07 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-NsC63dA01g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f261387a9d31ddfd8e9527d4b74b69e20bca0c4a37aea3ba713ae5f94127278f +size 114479 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-NsCNnryC9c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-NsCNnryC9c.flac new file mode 100644 index 0000000000000000000000000000000000000000..b6ecdbc72e0c93b09f9c2c0f408f672278dcccf8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-NsCNnryC9c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:899c1c4410eee3b85a270afeb9417293e4e1228367879327beedfcf76650a470 +size 132458 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-Qkeiyhg2zE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-Qkeiyhg2zE.flac new file mode 100644 index 0000000000000000000000000000000000000000..d8181b18fd1ebbb0dd0838aa2755b7285e0a7165 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-Qkeiyhg2zE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5c938fd1ad40e563175d7714d4b469c3a6229dad4eea704b40e189044cb74cb +size 151076 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-S4jm_6mb3s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-S4jm_6mb3s.flac new file mode 100644 index 0000000000000000000000000000000000000000..572c80fcba0ee239f1d1a35b221725e2845f5c82 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-S4jm_6mb3s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a38079870a4d42962f1cfdd952e7bc8d4acd879d5b68b44269ee47f30fb5ca6d +size 110570 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-S8UABOd2R8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-S8UABOd2R8.flac new file mode 100644 index 0000000000000000000000000000000000000000..901b0ecd40c1dc9427c074f3ac17ee73a0239364 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-S8UABOd2R8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:776133854e5ae335b1c5f61a8cad0a2cc8a9cf39abbb86612e04c3546e92f021 +size 259554 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-SuEOOhyRKs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-SuEOOhyRKs.flac new file mode 100644 index 0000000000000000000000000000000000000000..1384daeb7ea91be65ede6bc85dabc7b0cfe387ac --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-SuEOOhyRKs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8017b02c71e45061980148a303a3e69a4a224fa95ce7544a16c3f211093fc120 +size 171279 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-T1RU8oThuQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-T1RU8oThuQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..c0a56e53260f858f58ca5cb1353cced87544a725 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-T1RU8oThuQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13d035d8044b7a732a9cff43b4325be33b9a7b53c2306a713384e6bf606f84ce +size 258421 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-VCN1giSq7c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-VCN1giSq7c.flac new file mode 100644 index 0000000000000000000000000000000000000000..3079dac373a57a08b242a37808ee427f1dcbe34b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-VCN1giSq7c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd8fd5b3aff35a2628f921d54ab98bc0c89c74d638d748f9ffb45044df48e478 +size 215888 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-WE3eMj91tI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-WE3eMj91tI.flac new file mode 100644 index 0000000000000000000000000000000000000000..d370acf01bf2593f0df0be22f07b8936be1bc1a8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-WE3eMj91tI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de556e1dbb49a49be38d2c194e0bc087c6ca6858eee82a805fdf4968e3dc10a9 +size 157739 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-WFcBNcMBMQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-WFcBNcMBMQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..d10525d7b0965d638f1a1d0fd92fc81fc11abbb8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-WFcBNcMBMQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de9899baca8843fc513b02eec8a5891e24f73f55118d33e83bd1a44e29fc33e7 +size 228084 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-XGsWplBxsk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-XGsWplBxsk.flac new file mode 100644 index 0000000000000000000000000000000000000000..8eb02f0f1a9a482bec9c1781cb11e6d8f1087803 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-XGsWplBxsk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f77a52cfa7f3ad698bfe643ba9501ddd8cb205670e462886adff43ec2da7a692 +size 254684 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-XWE0n4zGbU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-XWE0n4zGbU.flac new file mode 100644 index 0000000000000000000000000000000000000000..0a6667d3a5b86486a55942776708105d1c9fbeb7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-XWE0n4zGbU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4413dee9ea0a549c601b35c6af3d41a828e4ff7b9239db8e12e2db2b26b85ee6 +size 197428 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-XZsrkx1QTg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-XZsrkx1QTg.flac new file mode 100644 index 0000000000000000000000000000000000000000..2e5cfd16e17e98be5756e5bd7c752249a2f32084 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-XZsrkx1QTg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a59f60f6285a64657d1ec0b69a7da314f099215e397dbd3c148ce147a40f15ff +size 234731 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-YDgnw597_g.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-YDgnw597_g.flac new file mode 100644 index 0000000000000000000000000000000000000000..3a436527645dd874d4291e8e4a939d04b5f587be --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-YDgnw597_g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdda87c19fe5171e0813856892567dba0750c37efdd6d96807d0a911c78385e3 +size 233451 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-_iFkUjP1Og.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-_iFkUjP1Og.flac new file mode 100644 index 0000000000000000000000000000000000000000..0e76cd28d5d4b1490373847ed5f8bfd87db2a646 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-_iFkUjP1Og.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b78dae541cafd7e84fd0b3587ed06a9013983490dabde66f1fa241a11cd69d +size 154987 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-a9HMUYHpSo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-a9HMUYHpSo.flac new file mode 100644 index 0000000000000000000000000000000000000000..44b5fda20f1ee17eaec5029aa1b0ccdcc0590b47 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-a9HMUYHpSo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88c8613e36f6360f7cdacc5d8f3d60b917a3369eb8753cb3dda2d37a1881bcf3 +size 246997 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-dDpGRPHzM0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-dDpGRPHzM0.flac new file mode 100644 index 0000000000000000000000000000000000000000..0c6c806eddd4a65693c758132b4c1b97077ec57b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-dDpGRPHzM0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a53b29d5b472218b2632969acc77ded37afe0375b9b7247343936c88863f9c0 +size 213866 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-dpI6JsEyLs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-dpI6JsEyLs.flac new file mode 100644 index 0000000000000000000000000000000000000000..4086b4519d2c65999c06d162f71ca06bf2423497 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-dpI6JsEyLs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dbfaa5df93783f90803eb3593647fbbb0f7a1ea4e28411ff44143f836d43bcf +size 254110 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-fl-wtFzC44.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-fl-wtFzC44.flac new file mode 100644 index 0000000000000000000000000000000000000000..8b87493e02f5c68085c706713316b6d197181d4e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-fl-wtFzC44.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b76de0c2ea57fed9a09655e10f306441baee08e861b4f78965025a1ac867a16e +size 149815 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-gAYyzIl2-w.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-gAYyzIl2-w.flac new file mode 100644 index 0000000000000000000000000000000000000000..d3abd6c323ca1d9531604b98c02ff9e442ddec83 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-gAYyzIl2-w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fc5e13b0b41b4028233936859be013a5e4f1e7a22e4de97db1b0b2a2b53519d +size 283375 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-gT--kmfcqg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-gT--kmfcqg.flac new file mode 100644 index 0000000000000000000000000000000000000000..007cb11f683deb38282409c6b55e44d7d1e6d82b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-gT--kmfcqg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d2c2f9efb30dc4743ce389d1e84610ccd7e38a8a3027c99005a013badc6ec39 +size 303317 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-iYTyeANFVM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-iYTyeANFVM.flac new file mode 100644 index 0000000000000000000000000000000000000000..307ae832251c2c869b850dbbd4e45c3107d5dd1c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-iYTyeANFVM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7473c74afc93162ccf9fd71906bbca6097d83a91ec8dbd32a38c0a2a6f096fa +size 189710 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-jCrIH2sB0E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-jCrIH2sB0E.flac new file mode 100644 index 0000000000000000000000000000000000000000..a8ab28bc7ce7e0c7cb6d0e9d1d9085179bfcad33 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-jCrIH2sB0E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:176a5f3e3ea3db36444029e4df8129f48a7cff80e4c5072b97323ddb777f321a +size 179865 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-jCyCT3KjZA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-jCyCT3KjZA.flac new file mode 100644 index 0000000000000000000000000000000000000000..d8c7f41600e656ef165b54f5ea37207ef2479aa2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-jCyCT3KjZA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cc91a8a4dcdf0879491aaa08b51c5fa79314116b015c665e96ea729a2dbbf7f +size 158707 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-mb4Fw4Z0xg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-mb4Fw4Z0xg.flac new file mode 100644 index 0000000000000000000000000000000000000000..cd84bfbb2bc7827b3dac8cc995dd57052b043f2d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-mb4Fw4Z0xg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43707dc437db70d9598a549019b6d88294a2ec72793e5faf60476586986a8a86 +size 239956 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-nJOtEC0tks.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-nJOtEC0tks.flac new file mode 100644 index 0000000000000000000000000000000000000000..876d91c51bc4508ca83a438fb58c3e0f4aa6e642 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-nJOtEC0tks.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d1b9d334e6b3124fe386a90e25a8c754d4878162f4047515004cf1adfa65b38 +size 224415 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-oy0BkpMGAk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-oy0BkpMGAk.flac new file mode 100644 index 0000000000000000000000000000000000000000..a6a6f54b70e75ea53a0f2f68034a47be120796e6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-oy0BkpMGAk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e594ec16d12d4b6472526ca562f246e67884540f7a3815584d6b2f7a18fad55b +size 261249 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-pmqKob1EHs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-pmqKob1EHs.flac new file mode 100644 index 0000000000000000000000000000000000000000..ee763aaa43d177d3dfd9118aa2987197f6bdbe5b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-pmqKob1EHs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f9ffdcb5cbdd4f86ea5595ac2c7f88bb91c19a94bc3153f6a4cd612b4381680 +size 238268 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-q3fQSq7Cns.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-q3fQSq7Cns.flac new file mode 100644 index 0000000000000000000000000000000000000000..4acb1584656eebe07c6fdd199aff34a7e5c3a38d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-q3fQSq7Cns.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4422ee038c28842cab21f14ae0efcd461c575372b386200a8d7429f0a8d15caa +size 174659 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-ttKN5Ba368.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-ttKN5Ba368.flac new file mode 100644 index 0000000000000000000000000000000000000000..79bc4c0b716bc27f3b2d4ee02413705833ca950d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-ttKN5Ba368.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:659b663ba262ef20ea7ee3aaeb23ebb1fb272510219d1069d67043b1e73e2442 +size 213569 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-vAJomuQQsw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-vAJomuQQsw.flac new file mode 100644 index 0000000000000000000000000000000000000000..8141e950892024806d725ad2190c4cdc9c33da00 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-vAJomuQQsw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f91f17d312c8814d2f546f8731bdf5d111fd0b0f2b3dd32732e77fd821bc10c3 +size 258276 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-wDemwOuiuM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-wDemwOuiuM.flac new file mode 100644 index 0000000000000000000000000000000000000000..a444625ead4b99205ac85c0ee5a22bf21ed3708d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-wDemwOuiuM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dc54753afd7e6799e258ab9374d9207d57b620da3f40d22dd702126f58dd14e +size 210514 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-xieOaxGx8E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-xieOaxGx8E.flac new file mode 100644 index 0000000000000000000000000000000000000000..c66391903e4ca1b4d5fad15a70d6be6a59227771 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-xieOaxGx8E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03a7f30b416b98bba5abc67d624cbb878624802dc056f87f6b7157d4b4319e34 +size 186910 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-ybHTybEnh8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-ybHTybEnh8.flac new file mode 100644 index 0000000000000000000000000000000000000000..46286ad0a50987afbc1876ee9501e6e87be79927 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-ybHTybEnh8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:356223682974085a90722ef90cd97c81d96b06982b560e3dc68434c5377d18ac +size 160156 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y-zXeN78-kV4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y-zXeN78-kV4.flac new file mode 100644 index 0000000000000000000000000000000000000000..b2d45fa276011c864c3293ea1370a29d2449081e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y-zXeN78-kV4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6384e9e7d35fc336f94421e6b3ada1b6c4caac9084d614f19e0dbf7f45dbd1a +size 253567 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y00PMurXGTMc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y00PMurXGTMc.flac new file mode 100644 index 0000000000000000000000000000000000000000..3918099b158cbcf4e2a5dfe448e0f137b6d47c67 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y00PMurXGTMc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eaa88e5f355d5a1b38e5bbae50cb0fa8fbbefc336ee7fa5a51d91504dad87497 +size 214945 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y025yQh3lGLI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y025yQh3lGLI.flac new file mode 100644 index 0000000000000000000000000000000000000000..672fc9fad4fa71e2cc02eb5137a03930d1280480 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y025yQh3lGLI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b780861fdd32e30219894b3d08f5c0838529464f5de3650be1d9bdda90c7af06 +size 205691 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y02yX3kyPZqU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y02yX3kyPZqU.flac new file mode 100644 index 0000000000000000000000000000000000000000..e98e7900622d035ab9debd7e38db65edca4d655e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y02yX3kyPZqU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4310173b3bba178821f8c2563313da1042dbf274ff90a50b8098befb4cb6a251 +size 215614 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y05zALRKaib0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y05zALRKaib0.flac new file mode 100644 index 0000000000000000000000000000000000000000..5773b84ad72808be6c58c91768a56a97376018ea --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y05zALRKaib0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0da54e276e5b5e6342f31bc2e7a51f58515c1c5d4cf58a1b5ceadbe7bc08f24 +size 237161 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0632OqvXrwg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0632OqvXrwg.flac new file mode 100644 index 0000000000000000000000000000000000000000..ad9af90c4025685cb5a857ac38c2611088b3d42f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0632OqvXrwg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62e945b5eba038991d7ce33622e1d32a0305c2e6360fd21e22ada6142a5b82c8 +size 131050 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y06seVx9COw8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y06seVx9COw8.flac new file mode 100644 index 0000000000000000000000000000000000000000..f873f8c51aa1ad789574e1212772969119116a13 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y06seVx9COw8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f1c26de9f2623993d2087d2bbd08dafe319afe41f681bb6de22705d3df1dbd7 +size 199924 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y079vcq6drhg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y079vcq6drhg.flac new file mode 100644 index 0000000000000000000000000000000000000000..b1cb9ae3111e21fe979c93c60543dc8f8dca3be2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y079vcq6drhg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31f0e20b199cdb2cb0f4954b0d74cb3bf28e71b919b351d5e233674443bbc185 +size 254679 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y089a2UWj08U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y089a2UWj08U.flac new file mode 100644 index 0000000000000000000000000000000000000000..08c7b7ccb8d3a556700d42a7e0eb2dc493fe243d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y089a2UWj08U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c895c3484af00cc05bec3a63590b502c4a494216a7f037f0172f86f673e92249 +size 221121 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y09hPW_9Q8Rw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y09hPW_9Q8Rw.flac new file mode 100644 index 0000000000000000000000000000000000000000..01cb2d4f05af5b1905879c3c3fe1749803f34d53 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y09hPW_9Q8Rw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6eec01b3d1544e92383573e6c27c7b0d53236649c9fe503f2dec6b751702b28e +size 230802 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0AuWiT7RBUg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0AuWiT7RBUg.flac new file mode 100644 index 0000000000000000000000000000000000000000..026d0f8a5932a9e419342cfca2e632ee580d1a76 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0AuWiT7RBUg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:281be6cb43335d030f0483328a7a822686b88987911ccd51ebc5dbb69078fbc6 +size 234177 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0Bdw87rcOdg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0Bdw87rcOdg.flac new file mode 100644 index 0000000000000000000000000000000000000000..d9b6ce288afd37b4e8e09c8bcd1731c1b8a900ed --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0Bdw87rcOdg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49f516f1a3628a707ec08e049d4df57e830153fc648fcef4a04d3fdad1636aa4 +size 251524 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0BwgEjw2aCs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0BwgEjw2aCs.flac new file mode 100644 index 0000000000000000000000000000000000000000..4c8923381c780bae1de5a569fb85ac25d20fa4b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0BwgEjw2aCs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19092f295055a95fb8e1310b6fcc95ec2eb6501732d3e1a464e2a41d756f8e9f +size 257097 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0DL92_Gd79M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0DL92_Gd79M.flac new file mode 100644 index 0000000000000000000000000000000000000000..df0c8fb1b4ee875c21c8e70d67e4f4b7228e6e57 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0DL92_Gd79M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e26dec11b572c6cb3951ab294830d50a8547f7175f70dbc08028f492a63bfbef +size 210399 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0EVyUEVloBc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0EVyUEVloBc.flac new file mode 100644 index 0000000000000000000000000000000000000000..bcfd371654f5c115e5e24a8ee42dbbcd643c42ef --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0EVyUEVloBc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16a2a3244993af75fd07a329aab42eceec3110c462b70519991e1a892dbbec1e +size 64576 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0FeuJL-r8m4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0FeuJL-r8m4.flac new file mode 100644 index 0000000000000000000000000000000000000000..358a7467cc3603f14d7621a77146af970ce54a5d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0FeuJL-r8m4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b88a6bd39458ce8f13336adf1e1a8ecc74a8e514152bf73119bf8de72bc086e9 +size 175788 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0G-Qx3Gv01s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0G-Qx3Gv01s.flac new file mode 100644 index 0000000000000000000000000000000000000000..7891712e3486b45705c93da0b0ebb244ad56fbf9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0G-Qx3Gv01s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a8816999c533a5f289c419bcc12135199b253721147519b1e69bc61f4b71fa5 +size 214437 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0GJCbGp_v1g.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0GJCbGp_v1g.flac new file mode 100644 index 0000000000000000000000000000000000000000..60027efe0d004f66a8c3d839a1bb03ff4141e93f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0GJCbGp_v1g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:339764299b7c135deff4a80644daa21a7de404fef3471c60b7689e2ac4ebe5f7 +size 200851 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0GRzxYNqvjI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0GRzxYNqvjI.flac new file mode 100644 index 0000000000000000000000000000000000000000..e88b6e16f43e887d3487331646b383af72acc0f0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0GRzxYNqvjI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93489720bb4d7bff7cb850a848321cacf276c1431b31b469d80da37a34bb6f5d +size 251114 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0HDD9x7hsZ8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0HDD9x7hsZ8.flac new file mode 100644 index 0000000000000000000000000000000000000000..6c72d3519029254eb90e73888bbd6746fe9ed0ad --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0HDD9x7hsZ8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6442645cd1bea555da115c63e8e16b19c6c9653ec58c13387c39d3cda77d9a66 +size 281428 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0HMa_BIbTo0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0HMa_BIbTo0.flac new file mode 100644 index 0000000000000000000000000000000000000000..a6bf9b44d77c393c85001ac631c57c0d39e20de6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0HMa_BIbTo0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3b718c042b7228faed55dabe955613fb7386bc9ce01d402d2f1ec900f90d8bc +size 208487 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0HRItbIdDak.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0HRItbIdDak.flac new file mode 100644 index 0000000000000000000000000000000000000000..6f521498d613a46065a2f0229fddfb5827d51254 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0HRItbIdDak.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:574b8328d5f5e9335606d17bbc2e6d3b393c17e619c31b758ef7f7c8ad250144 +size 272822 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0IGsVL49I2M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0IGsVL49I2M.flac new file mode 100644 index 0000000000000000000000000000000000000000..218facafbc733ce2e5c8b525d1f74c8e7965078f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0IGsVL49I2M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b2530c1421f32afcf53d224830a2723433da847b611501088fb09f75ccc62cd +size 127008 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0IGvw7Eplrg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0IGvw7Eplrg.flac new file mode 100644 index 0000000000000000000000000000000000000000..3034639f20588b49033e6e712a076c5a312fded7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0IGvw7Eplrg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:46ffa16657f26e451d05f61e7fb4bdfde01afc5dd7d58453603386ac425e689a +size 213570 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0JqQzE4wGEo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0JqQzE4wGEo.flac new file mode 100644 index 0000000000000000000000000000000000000000..5d0d710f49da8799368bc5f74c6f48f5054b1959 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0JqQzE4wGEo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1f416d33260f8dec6deb3b063dceda60d82fab247769fd548915eba89df3d0a +size 198213 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0LezpTtLGik.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0LezpTtLGik.flac new file mode 100644 index 0000000000000000000000000000000000000000..74fb4ff8eec7ea7b0f96a1aa252ed3861a33416b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0LezpTtLGik.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cda63e3f37e5290d2eff69aa31b16d90dbd5e55254ef74af54b172069fd4e7f +size 263571 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0N8udwD-wXE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0N8udwD-wXE.flac new file mode 100644 index 0000000000000000000000000000000000000000..ed0e68f8f6b4bd229733902a1cdb857141d6b3b1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0N8udwD-wXE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4292a273a193d24aae6471135a7e8213e9606f607cd7144ce2f04747b1d21c0d +size 252333 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0NvU5eiSlJw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0NvU5eiSlJw.flac new file mode 100644 index 0000000000000000000000000000000000000000..386f8f94e1d6577e342d98f0ce5d0e3af1ed412b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0NvU5eiSlJw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9262ff6edd11ad6261cd7740db720cbb6fd1fdea978b86141e2737d0a9874cbe +size 194177 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0PJ6PRTzoWs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0PJ6PRTzoWs.flac new file mode 100644 index 0000000000000000000000000000000000000000..4b56ee7229804d475104fa94671f4648ceb4633f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0PJ6PRTzoWs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0022637ad1ab40f47fcb741f3a4736ea121a41db2c8758ff9db790a5472b7d3c +size 261558 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0QSg9bSOgbA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0QSg9bSOgbA.flac new file mode 100644 index 0000000000000000000000000000000000000000..3b46f4dc518301e2129fccf366b6beec204431d6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0QSg9bSOgbA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8919a9df094d4089ca10ee9a591db8b5f58b311e98c22962477a7ce890e4278d +size 181380 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0T0MFoavHic.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0T0MFoavHic.flac new file mode 100644 index 0000000000000000000000000000000000000000..fcf085a6be4c48a2dedb224f59d3998497f41194 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0T0MFoavHic.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6bc6d8e7b48076acdca4d5f9662880d2bf7c24768d4f36f1e3b2a77f0ae9b2f6 +size 183988 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0TumqT-_n9k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0TumqT-_n9k.flac new file mode 100644 index 0000000000000000000000000000000000000000..c573303bf3fec43f06517e47c5c49d4d3e113c0d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0TumqT-_n9k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db76b3785eaf830bd6211a06a55eb4c339d74e7b032a83754b1667c0b3f9c831 +size 248218 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0VlXTRYdzcU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0VlXTRYdzcU.flac new file mode 100644 index 0000000000000000000000000000000000000000..fccc3aa1b230abcc2b06827981e15eed31f7a710 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0VlXTRYdzcU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fce03839104e0ff518da4755efc9999353018f69a0b1faee501b02025e18f828 +size 200329 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0ZIy0COr3NM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0ZIy0COr3NM.flac new file mode 100644 index 0000000000000000000000000000000000000000..36722c088414fb0a10fafed606f4e26b2a10ebea --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0ZIy0COr3NM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a3b65021ff01d759be3299ce6fa82cbc1f8d27af7b28e9df4345b1382d613e5 +size 201136 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0_zLL39nfoM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0_zLL39nfoM.flac new file mode 100644 index 0000000000000000000000000000000000000000..78df037e7ec890a2c4922635d263a08b6d961e03 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0_zLL39nfoM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1b5dbb2d744124baa59be9d7ecef21e9afb0d681eb18732ba991d3d60f3391b +size 245721 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0dVfocjWM28.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0dVfocjWM28.flac new file mode 100644 index 0000000000000000000000000000000000000000..c8361b80f58e47794626666038eee6d3a4be61c1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0dVfocjWM28.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:691a33c6c3380602e09189cd8ef36507b6f86d35c39468bdd99a29f2ad0f7f76 +size 176701 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0fECwjwdQUw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0fECwjwdQUw.flac new file mode 100644 index 0000000000000000000000000000000000000000..db6c9ac8c03c3114ca8d8669865c114c3e35c7c9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0fECwjwdQUw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65d699eef10693694db66cc6a862a7e712c80ab506a73eaec6b7f29f7dc0f41d +size 217909 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0fEnGyWYLCc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0fEnGyWYLCc.flac new file mode 100644 index 0000000000000000000000000000000000000000..c2240411d058b1547e0a4c60a81ec81bead5da3a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0fEnGyWYLCc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f4a816fd52510e78a99054b8aab6b316c76e5c9d4affaca7c8886e7110da9f5 +size 274001 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0fPqbvlhkOc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0fPqbvlhkOc.flac new file mode 100644 index 0000000000000000000000000000000000000000..2c26609c6100edf324eb80637e8de0b707852a87 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0fPqbvlhkOc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b3b528aeb57ccfb61fe59189168af1c90f810f24513657c8cf7acf0eeb4c940 +size 213083 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0g1DCi18pZA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0g1DCi18pZA.flac new file mode 100644 index 0000000000000000000000000000000000000000..06e7e41b61e9522e6b9fbe00aed951241e6d897a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0g1DCi18pZA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:742649dd9bfbc7be9d70933b07678fda18d13ace69feba677c77c60fefc8f19a +size 133936 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0gD6-Tes60A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0gD6-Tes60A.flac new file mode 100644 index 0000000000000000000000000000000000000000..a08d73a2f4073a4c104c23c43d3a4ebdb00104f7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0gD6-Tes60A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80597b6b18bd52d1cc806109742080b0205e558612c21a4f4f86dda56a4a6a71 +size 227998 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0gfxwS7Wla8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0gfxwS7Wla8.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7099d5da6a8c92bde0580beb70f1bab1fde2a72 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0gfxwS7Wla8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e1caf5c8169876a74bc0f7a30fbe6ac66ac558cd209a477bdaaa9d1d97e6053 +size 259305 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0hKEM6Tq6zA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0hKEM6Tq6zA.flac new file mode 100644 index 0000000000000000000000000000000000000000..022ac68589fe1fc89bd05e126789f4c51ba2767c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0hKEM6Tq6zA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04e5b390901217b9ad4e05814753fa41ccd48044ba3ecdda5929ec3f680c5bb2 +size 149002 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0jS_ViL5QmY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0jS_ViL5QmY.flac new file mode 100644 index 0000000000000000000000000000000000000000..35c4515376005797e6416bbe935ff30e6227405e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0jS_ViL5QmY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7de18dd4ebe7da04227d5980b5ea8d2c3afda5695ee5efb60ba1c169cb7a6fbb +size 236616 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0jTGXM7yEqs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0jTGXM7yEqs.flac new file mode 100644 index 0000000000000000000000000000000000000000..73551613ba6341876e6d5e0c715b14f5e03440c4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0jTGXM7yEqs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fef11011331d19a14d85564b79cdc7c9c732d12548a05019ddfec15d4b437f36 +size 230555 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0kueqaUidvk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0kueqaUidvk.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7c37cb135d1919130f391affda5129b5a436033 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0kueqaUidvk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0fa968691b1dc8981799a336000998236873cf2bf702400ccfeb2293f486d91 +size 254989 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0lQ8ZEfBKlI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0lQ8ZEfBKlI.flac new file mode 100644 index 0000000000000000000000000000000000000000..c7084506ca51e1cf9e8564c81ec1a476c477df4f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0lQ8ZEfBKlI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af076479f2ddb7b8fffcbf3bfc365060af0c56a5ae370a30ee28daf7dc4aee54 +size 171309 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0nXHO9hRfO4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0nXHO9hRfO4.flac new file mode 100644 index 0000000000000000000000000000000000000000..4fe359224af5e21d46970bcf35faae31c73f410e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0nXHO9hRfO4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f6f7f312d805882241926bce774c1b3b030d8e4ec2c00035f62a95234efc62b +size 216012 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0oKbg1lSPuU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0oKbg1lSPuU.flac new file mode 100644 index 0000000000000000000000000000000000000000..68f792d678081d33e659f4fb4b8f19772e33b662 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0oKbg1lSPuU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ff7c5f25368912de1a125e43751d54d4548c8c1461d1a56abd208e36de58f9f +size 221601 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0ojTL4MmXjQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0ojTL4MmXjQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..8bfd3e2385be30e3c583e820de01529c3f90c388 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0ojTL4MmXjQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7f216d459f7ee4ed4a63b007c5f09716417aa2d8b260ad3ac1bf6482e0645d1 +size 237612 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0pNLmwA8DdI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0pNLmwA8DdI.flac new file mode 100644 index 0000000000000000000000000000000000000000..40506f6afa8aa5807b17680e9f87b152f26f783e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0pNLmwA8DdI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:043ee3377432a4346630284ad5fce064d0521513119e6b53da97d38ffb711782 +size 236021 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0pvQ0fpTmW8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0pvQ0fpTmW8.flac new file mode 100644 index 0000000000000000000000000000000000000000..b46424754ee3eff679cd767438d0c961cdb5d568 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0pvQ0fpTmW8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:483aedd2834f11184f740197f75673915044c8e9e4f9fadb2091a6c1510ff0c1 +size 235612 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0rcAa1-6z28.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0rcAa1-6z28.flac new file mode 100644 index 0000000000000000000000000000000000000000..6799af17ea10be895d469a4243d5e680bb4c45ef --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0rcAa1-6z28.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5432b12fdf811e183519c5ac20c5e187247e5f06b0c43f027e61a129ca60213 +size 237900 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0uRkdMtCuDM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0uRkdMtCuDM.flac new file mode 100644 index 0000000000000000000000000000000000000000..dd91be4e5ad0fee11b98c01dc66297b17555414a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0uRkdMtCuDM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4764a386f4f1be253e633233d2d08e1ee8c644433d3b1b7df3c70a26664df617 +size 232810 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0uab4-3d6MM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0uab4-3d6MM.flac new file mode 100644 index 0000000000000000000000000000000000000000..4529ab2a4eaeb236ce0d8be03fc03bb3790d9a9e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0uab4-3d6MM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7edfdb8dd918349bc04f28e214f574a1d134d1aae3e13d7845f6836a3d4c408a +size 180218 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0vOgLGk1dt0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0vOgLGk1dt0.flac new file mode 100644 index 0000000000000000000000000000000000000000..3f0c56bd5aaa9bd0882a974616e8118cfde3e5fc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0vOgLGk1dt0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da498a4c16562ff2530c1f185d10f025d18cf5dbc7b3392f52890a6342cedb11 +size 181909 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0wgFRL-xcd0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0wgFRL-xcd0.flac new file mode 100644 index 0000000000000000000000000000000000000000..7914576385fca9756f700c88de70bbc7ef19f7f3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0wgFRL-xcd0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f81b290f2348c1a4fb80444135e5d2bed756ac7566851dd12b007fcb6e4b72be +size 191142 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0xIwZLID2i4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0xIwZLID2i4.flac new file mode 100644 index 0000000000000000000000000000000000000000..443dad603831223a8f9e471a6b6caf3fe3544c67 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0xIwZLID2i4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b4bb92e1d87bd5a519dca08804a5e5792061fa91a189897ebe231267963785f +size 271414 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0yETgW44MZU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0yETgW44MZU.flac new file mode 100644 index 0000000000000000000000000000000000000000..5f9794626cd6b9f3a5f0a74db3e1a89884cfc0b7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0yETgW44MZU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0714c65252794121415065c45aac074f5fe29ef5aa2de49e3c6d16f3c79e67d6 +size 244521 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0yxEvdnimGg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0yxEvdnimGg.flac new file mode 100644 index 0000000000000000000000000000000000000000..49c88c932e3ff0f7b02b79d25036a7ae9fabc12a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0yxEvdnimGg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3b187830d9bb9c53941524c1dc99422324a1fdba14498d92da6eaf9c937ec1c +size 191217 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0zOfqWQuies.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0zOfqWQuies.flac new file mode 100644 index 0000000000000000000000000000000000000000..01c2b11b604bf36c5e7f4ff5395220f5c6b27288 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0zOfqWQuies.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b56e428eb558153a2fe058720547cb4a2660ae56fb57c2d94a6398e115c9461e +size 209301 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y0zmN7Lvm_J0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y0zmN7Lvm_J0.flac new file mode 100644 index 0000000000000000000000000000000000000000..f2e1f2bc0d3222a29e5486feda30b6cb957a91c6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y0zmN7Lvm_J0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2ed9f6cf3e043693f88cf79813c211531c0e12d019674bb47081e7806fe54f2 +size 160334 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1-jv-736aIo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1-jv-736aIo.flac new file mode 100644 index 0000000000000000000000000000000000000000..373e1fabd180439e314f412232dd87df3119bd15 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1-jv-736aIo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:246c4761803806ec4065cdb0de7df68d4df828250e70a83b93091b965a2abaa4 +size 216478 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y11cctCD0ATQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y11cctCD0ATQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..8bd79fb051c810e818a8a3d5cc88c883d043c383 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y11cctCD0ATQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1c78738864d15c1418122a86840a4b15b8727cf51c5b56cc6126b0b908c1cf4 +size 188136 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y11qcdPN80x8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y11qcdPN80x8.flac new file mode 100644 index 0000000000000000000000000000000000000000..800929aaaa9383c564e867a1f153e480d3f4b770 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y11qcdPN80x8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c80ba6cb674c0d207a54273a368be6541cc438c32e7211a3b9d1c446d50bb7f4 +size 126529 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y13QUBj_VlIc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y13QUBj_VlIc.flac new file mode 100644 index 0000000000000000000000000000000000000000..fbd207ec24363f16570fe8e0f635d20daf5e972b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y13QUBj_VlIc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c32d090209afb001792ebde623d23bda3385df8dc95ca4f234c3098fd60eef7 +size 184191 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y13WnlxzBU6k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y13WnlxzBU6k.flac new file mode 100644 index 0000000000000000000000000000000000000000..0e1caabac10370d7502f62b6e1529139e35c794c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y13WnlxzBU6k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0299103a32d03625c0410f328db2374f91002f73ed7ec5e66fd7f7653cabddcd +size 265428 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y164fV2R8E4Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y164fV2R8E4Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..1e683f92f27d4068c52ad781628001cbbe0955f7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y164fV2R8E4Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27e432bf9e30e91ff0b977df8734155bf688a5006febffea315eb9473378a2c6 +size 238343 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y165obDPbUVw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y165obDPbUVw.flac new file mode 100644 index 0000000000000000000000000000000000000000..6491990f1324ffaa422cb7cf934c6b849c5cedf4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y165obDPbUVw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64a8442c86ea2cd3ee21717bd58fad4843db93fd66bc8f35e088aa375f4cd865 +size 166830 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y16KqWYIlR0s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y16KqWYIlR0s.flac new file mode 100644 index 0000000000000000000000000000000000000000..0d18809c82f724bb08dcfd604d0efad5b7bb0ca7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y16KqWYIlR0s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58f3b5222e4ec6da24830e9fe448e6a6180b86585c9d950d668e03761a6a4b05 +size 217727 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y18FN3fDI25A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y18FN3fDI25A.flac new file mode 100644 index 0000000000000000000000000000000000000000..3ee610de2c8d6fcf4fc182ec3f1aac4c946f3df7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y18FN3fDI25A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:198228abfe0f48578a02abc249c39789b2630bb90c491dde8f2be54873b3e20b +size 209010 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y19rWZojaOK8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y19rWZojaOK8.flac new file mode 100644 index 0000000000000000000000000000000000000000..f580573252277d6dcfead04a418c4e812df0f786 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y19rWZojaOK8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90f66c2f6654ca0ee5b81678ac1782d94938031b4053fb097c13c1baa4c57bf9 +size 217941 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1ATOT2TZ5Gg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1ATOT2TZ5Gg.flac new file mode 100644 index 0000000000000000000000000000000000000000..b071a8dff2421f041f903f361b9deb0b7947eb66 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1ATOT2TZ5Gg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a09ef1c5f47d23e9333239405727c095601522299f43836a52777539cfd12b3 +size 185918 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1AejlUdd5gs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1AejlUdd5gs.flac new file mode 100644 index 0000000000000000000000000000000000000000..a8d48d60ea12f72d93b308bd578cf6740dc674f0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1AejlUdd5gs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6e4a89556d3d605862dad4a811c9408326e1882a5fb65f6fb03620da152c2c22 +size 150010 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1AjYKvdjOJM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1AjYKvdjOJM.flac new file mode 100644 index 0000000000000000000000000000000000000000..09167e4dce2916dafdb0191171909f418bf2e3fa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1AjYKvdjOJM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8211e37d0635342540337eda2bd25f8483205a809d589d57edaf2f65566344b3 +size 163877 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1F7JD5A9iTI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1F7JD5A9iTI.flac new file mode 100644 index 0000000000000000000000000000000000000000..e306fa4b21f6f0179e50a7389d8d6f9b5b231d83 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1F7JD5A9iTI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:facd5162228a89c642d0bd0803fcafdea6d8dee33ad58bc1a19f8c768b856db2 +size 166064 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1Id9Ow6_v1M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1Id9Ow6_v1M.flac new file mode 100644 index 0000000000000000000000000000000000000000..0a3a0aedc076b0103e16653d3ee8ba4b953f422d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1Id9Ow6_v1M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45969d38fe03e525cd3cc29e326430b545e31cf95ecaaddb6527403bf6ecd9de +size 227186 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1J-SyJKJZa0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1J-SyJKJZa0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d9f12afcb850e7dab1f76977c1dba5ddc138cd52 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1J-SyJKJZa0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c357f9c2a5c409e1797df26a9074945a90985aff280898a2aa9628d48eff945c +size 217023 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1LGQ1dUHT0Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1LGQ1dUHT0Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..6b8b590bf1bd1c37a299d80067a7726ca94bb15f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1LGQ1dUHT0Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6ee86226f58d4c664c07915d4a34e8e6d1ad2229d079edb654eb1b8e079417b +size 216010 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1Mc63idyMZA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1Mc63idyMZA.flac new file mode 100644 index 0000000000000000000000000000000000000000..6e50332cabe3fbcad67f675828de4b29d7b902a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1Mc63idyMZA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a5f7fe35d38dec80077a9dfef8f2f46a840af268192abf619dcf4f566db8ee3 +size 290119 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1NVfl7AnSRg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1NVfl7AnSRg.flac new file mode 100644 index 0000000000000000000000000000000000000000..274295c3c7ceeccf8434cb180b6a157f63135e33 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1NVfl7AnSRg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2731c7825ea32f1b7fe532f681a9cd2973e7534557c2f147589a15edf352169f +size 215718 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1Nl3caLibno.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1Nl3caLibno.flac new file mode 100644 index 0000000000000000000000000000000000000000..00e51a53bddd277b5d6ded38af3f02d41c4f477a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1Nl3caLibno.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9583087885b5cf9dae2202293c1d3e8e23714ffdeec5b0542db9ebed73d74d12 +size 253520 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1NrW4c8fleU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1NrW4c8fleU.flac new file mode 100644 index 0000000000000000000000000000000000000000..d2863afcc356806a1b10c531e4857f2ef5359ec3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1NrW4c8fleU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2006e27263658f977a7bc25c8af98fdee6f0e902d5ec67a437bc033333409741 +size 264042 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1PMqP8xaMbk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1PMqP8xaMbk.flac new file mode 100644 index 0000000000000000000000000000000000000000..9a629573a593636d0b64509257a42376b7fc324d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1PMqP8xaMbk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c477af15ddf5ea6b4ba4a0499c5fe2dbf2b8f18565b611bc271e88115ae11c6 +size 148339 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1QR8acH95Ho.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1QR8acH95Ho.flac new file mode 100644 index 0000000000000000000000000000000000000000..4ca6f3e7f739a8f6b174ad4f71d8d6ead76ba09b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1QR8acH95Ho.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:037180749e23f73cc1662a2fa1259629ed7300de97ce5478e71cb6229523e6d5 +size 213450 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1R14_3kTmbs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1R14_3kTmbs.flac new file mode 100644 index 0000000000000000000000000000000000000000..9af953eae12d425d3caa37ec3b7d5ff383fceb9a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1R14_3kTmbs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54cf7224a756def185c01b57ff068905b9cfd0b73ddf9f2e8eb004058770053c +size 182453 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1RpnBTT-dxU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1RpnBTT-dxU.flac new file mode 100644 index 0000000000000000000000000000000000000000..4d730659502fd1577e8b36020c403542d0fb8508 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1RpnBTT-dxU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04c0b2fbb3bbd7b6aa2ecc94563b1e45cc3a32ded2f425f1741ae9f3bf45044c +size 251380 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1TIx22Pnt1Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1TIx22Pnt1Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..427e3a7369e17a12bc89b0168bbca523ee8588bf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1TIx22Pnt1Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58c1bd03c1ea2fb53b1c95099a9cb08c67f63420aa2324938e6feea268fbb8f3 +size 205652 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1U1k-d1x1qw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1U1k-d1x1qw.flac new file mode 100644 index 0000000000000000000000000000000000000000..3d271474f26fb41b7c5ed70c8dbf5baf2e4f3cf0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1U1k-d1x1qw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ceaea94c6745bbe64cfcfc3848ecb70064a2a61d60c369cc05d4231e89adb23 +size 120837 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1VQZAJgvzWI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1VQZAJgvzWI.flac new file mode 100644 index 0000000000000000000000000000000000000000..a0202b32a9e05a9e96f7886de54e78fb96b28aef --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1VQZAJgvzWI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24939518f8ba0e33dd3131f630d6fa8a5783fe9f00d29fba0bed39adc577c68a +size 181357 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1VVr6_KQGwc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1VVr6_KQGwc.flac new file mode 100644 index 0000000000000000000000000000000000000000..0154c820f87b2a67708f931ad58df4a26e3c2570 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1VVr6_KQGwc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea7cd6fd2c05ee399e454dc60bb7dbdcb7b2d02caf2b713138195fd1eea7c12e +size 281612 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1WzI9RX3CnE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1WzI9RX3CnE.flac new file mode 100644 index 0000000000000000000000000000000000000000..44af12ffef6f5283b8129151b44f36794d689e30 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1WzI9RX3CnE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be309659c85707e4d33c9d8b69c15fa38ae9a4ec5b7c61b50df7f9c261919cfa +size 258191 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1X2Haq3e8wo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1X2Haq3e8wo.flac new file mode 100644 index 0000000000000000000000000000000000000000..b1b6955bb847b068dee8eab677d6289f994d2e8f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1X2Haq3e8wo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:71eea7a35ae663745f81c97d269fc6f50ff2862f3630c7c3961d1c1a6d19dcba +size 291651 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1_bLIXzVyio.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1_bLIXzVyio.flac new file mode 100644 index 0000000000000000000000000000000000000000..06c9bedb6c0e9d21dc2b17ff8d7437a1df8f4e97 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1_bLIXzVyio.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0cedd9e224cb6462073c47a9b787a5f75f804c4c9f17d33376ad92f1754f259 +size 180049 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1b7WUzSzU28.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1b7WUzSzU28.flac new file mode 100644 index 0000000000000000000000000000000000000000..007d414b630dad670d097fc079b9bfa14ef6d407 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1b7WUzSzU28.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bbae9ef2af3aa16006bdc36c9b350e3f17694cf0cc9289f7f94c8ca2832e330 +size 251606 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1bAV_svnjUo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1bAV_svnjUo.flac new file mode 100644 index 0000000000000000000000000000000000000000..caea11a6a90b76e479f390b7641032553aa63147 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1bAV_svnjUo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7797a8c2187f2332a5da573919e27f1c3137a6f2410eee33a8387522f28834e7 +size 240635 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1bE_vX8Oy80.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1bE_vX8Oy80.flac new file mode 100644 index 0000000000000000000000000000000000000000..3def880e496382f8e1ff27e9b25ed31aad48d18e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1bE_vX8Oy80.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:034915dd74c64c2f4edf67a727d37f596c0764a94627bb11fc67c4677770266c +size 173022 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1dTDTUvU7yE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1dTDTUvU7yE.flac new file mode 100644 index 0000000000000000000000000000000000000000..4c13ec1017dd44b62cd2a1fe643daccc33ae8d13 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1dTDTUvU7yE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:156c44392809fbe75fc0b1091ab489fc5d88acb1c65c913f91f06b2474c1c43e +size 237557 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1dmz0KR31Xg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1dmz0KR31Xg.flac new file mode 100644 index 0000000000000000000000000000000000000000..4cb84c740e014bbc621a4cda207efedb74f81cb2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1dmz0KR31Xg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:40c83c093fc1ba04047e766d5f84c93adc660b814606e31d7bb4e8581da678c9 +size 160131 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1fwJTuGNNB0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1fwJTuGNNB0.flac new file mode 100644 index 0000000000000000000000000000000000000000..f7c8ccf1d09cde9fc7f40dfa6e290bc0d8e63aa6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1fwJTuGNNB0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7e1f12d66c24a1b5ab7f7f87b4ea16d75b84d8299a2c7d1f9e27952bf3213b5 +size 209924 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1iK9FP1ltuE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1iK9FP1ltuE.flac new file mode 100644 index 0000000000000000000000000000000000000000..073388c5df31f2bbdbb7c0732e796aa4ee388b04 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1iK9FP1ltuE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aeecabecbfcc2f9ab786e3a2ded5ea83a7ed28f9737d04bd431ef36ad0af223c +size 278804 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1iWFlLpixKU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1iWFlLpixKU.flac new file mode 100644 index 0000000000000000000000000000000000000000..69039dc3efff2b4dc8fbb1b0aed73a5f0951fec7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1iWFlLpixKU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37958ba42d3af318a07ef9c145aabcffefb2bb4904c6595cce7b2912289b64bd +size 179208 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1jD2iUXynFo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1jD2iUXynFo.flac new file mode 100644 index 0000000000000000000000000000000000000000..69f11f6a38ac35cb0eb1c1925b18493e53f48463 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1jD2iUXynFo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d10f273abff8991c3e83b8daa5f94fb207614262b79e31546a75edf62fa17874 +size 276405 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1kxnTHW60PQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1kxnTHW60PQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..921654bb528f354e6a5b4da5cb02f317550f32b9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1kxnTHW60PQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d64bea4c45cba6e98c4315327766862337095ed0e67fa774ba44bf0a4e5b68c +size 193394 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1otuMUIZ6As.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1otuMUIZ6As.flac new file mode 100644 index 0000000000000000000000000000000000000000..ab5fee7e37ffa31481f62adb2703871034bb37b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1otuMUIZ6As.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c7e9438407f5e554d6bdfef04f0abc1e5ef8e538ee3e96a55d58f0c3714f372 +size 195924 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1p1ekNTgCQ0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1p1ekNTgCQ0.flac new file mode 100644 index 0000000000000000000000000000000000000000..373e28c1356175b83ba8465878f5f85372c62031 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1p1ekNTgCQ0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c5123c23225eab5be2bc29a7bcb254cfab540128e6c06a66f6c36a84207cb5 +size 204991 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1p8YDM6gG6Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1p8YDM6gG6Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..bdf6cb5db0c6453cf19f7f29e930a58b23bdb323 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1p8YDM6gG6Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fcf180b95d46a0d0ca4b8e95130a3cad08a9306aa2da90e5c62aba6e5e40eb11 +size 181986 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1pDw9AoyZKk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1pDw9AoyZKk.flac new file mode 100644 index 0000000000000000000000000000000000000000..69e7498cf2f178ba127edabf1b82a4acee018a44 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1pDw9AoyZKk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5de7f3aa643d682b0fb41f1e5bc326681aa0ca3ead7a8a0b0bc6521f805e7463 +size 253346 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1pXuFVBomP8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1pXuFVBomP8.flac new file mode 100644 index 0000000000000000000000000000000000000000..2b228c815e657064f5f8d491414f77b7d5e67057 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1pXuFVBomP8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93b22fbe86966b3e5c6d0a42c5a24ac607b131852a29a4390b90072bdcce70c5 +size 263953 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1qWIWqV1gyI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1qWIWqV1gyI.flac new file mode 100644 index 0000000000000000000000000000000000000000..e6c518474ff9979333c447e2bfc171b5bed38f96 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1qWIWqV1gyI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2875e8bde75ab1630438eaa908b6e403d0d0153412d53710142cc3e9defa2a28 +size 232141 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1rPv2-OZLK4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1rPv2-OZLK4.flac new file mode 100644 index 0000000000000000000000000000000000000000..0c0b7fadf42d4673d2cbe5a753f87f2e6a842b80 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1rPv2-OZLK4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:802005e3ad2a9714b6a7607a71442fa35d9a50edc80d7e0a099d4841f1fba576 +size 225330 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1slvoNgzBLE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1slvoNgzBLE.flac new file mode 100644 index 0000000000000000000000000000000000000000..748d18b8786c42d4f52039feb44eb3aa837fb8d7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1slvoNgzBLE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdde99f2322c7faec2071dfdf86f19e7941182b487221b3af0371125a2ab2c88 +size 223562 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1tujDfcQZEs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1tujDfcQZEs.flac new file mode 100644 index 0000000000000000000000000000000000000000..4501b493985e43f732fdd84c09c444fd723096b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1tujDfcQZEs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e95a6f5a2187bc9fa422b619a8f63fcb261642157aa7310b05230071a2b7b321 +size 174890 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1vZNNVA7AMU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1vZNNVA7AMU.flac new file mode 100644 index 0000000000000000000000000000000000000000..898942ca25f1f8a1ba355e5f8679aee4b66179b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1vZNNVA7AMU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeec10daff385c93004b52158fa43b0d5b68db9e8c5ac6901935eb6033f8efd3 +size 218532 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1vmlc5ba-OQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1vmlc5ba-OQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..c69f533b6dbf135630cf26af1d0a94b55318ed59 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1vmlc5ba-OQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c562821eeb1d38606899d3e07aff740caafb027d29d59fe342e35315696a60a0 +size 240791 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1xTMKVBMU0Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1xTMKVBMU0Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..55d79c71bf7dc111319bec2f6e238c52b4350bae --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1xTMKVBMU0Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b33fb4ada4cf6ec86157285d3bf5e9295995cbd51f3a75ff911e3c9b0726d38e +size 290092 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1xwOwK2T55c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1xwOwK2T55c.flac new file mode 100644 index 0000000000000000000000000000000000000000..f505181fdc411401310b3950fd2aeada99effe7f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1xwOwK2T55c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3e0911e2f71df579f2d63cbc2ba85631b668ed460af96fc4b97086d0d53f5a9 +size 266186 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1y10mtbYQP8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1y10mtbYQP8.flac new file mode 100644 index 0000000000000000000000000000000000000000..eabcbebc001b92cd7bfea40013e35079078f285d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1y10mtbYQP8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:273bd811226b50e2064d4ce4fd527f95844912af54ef512404ef1eef867ecafc +size 215325 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1z08ytHRqfA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1z08ytHRqfA.flac new file mode 100644 index 0000000000000000000000000000000000000000..db4fd1bb925f4c060524db509c54e7d5aa1cbb34 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1z08ytHRqfA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa501ae6cbad954dc3c25587eb503c9f0cb1ee5489797ea9fe8b64109de313c6 +size 275677 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y1z_0FAzrbsU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y1z_0FAzrbsU.flac new file mode 100644 index 0000000000000000000000000000000000000000..b4c7de370a287537b236e80871c91fdfbbed6a28 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y1z_0FAzrbsU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83154371dcec1307a91e3f1551ea5568624e91e5f578acd55528c8b6883d27a0 +size 208240 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2-BcDQ7XO3Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2-BcDQ7XO3Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..14e3dc7e73aeafa9caf399e3f97bdedca447cabd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2-BcDQ7XO3Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ba49cb234f6ab193db1cb8643a77a94f16376e23ad9e5fb5b37d664d5cd52dc9 +size 219326 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y20RwAJ4dtSU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y20RwAJ4dtSU.flac new file mode 100644 index 0000000000000000000000000000000000000000..226517bf28c6f6467715d00d6b0feb7b962f7cad --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y20RwAJ4dtSU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5c4e4ff8bda76bf59f5e2908edd10283f080c254f23ae3b8d306a33b0cf5819 +size 184400 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y21WMz5hG0Zs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y21WMz5hG0Zs.flac new file mode 100644 index 0000000000000000000000000000000000000000..a2bc7cbfc27d067d3ace54526728a7e6b962df06 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y21WMz5hG0Zs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eec0e0396a0e159885fff467ff2e3e2bcf1a84cb5d17c0c806c47b2b62c64900 +size 239650 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y23mRMsHfek0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y23mRMsHfek0.flac new file mode 100644 index 0000000000000000000000000000000000000000..2676fe34683bbe5b043375f3fdb7a444096417f3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y23mRMsHfek0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbd1b92b20e26028fd3bbc079d56fa8b9265980ccd2586066ee0a5769f4e4188 +size 181443 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y24pz7PVwcw0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y24pz7PVwcw0.flac new file mode 100644 index 0000000000000000000000000000000000000000..5b5c1de8e1b4d773defffec6b00d7523b8991e31 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y24pz7PVwcw0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b15ceca4e99b8f189bec47f5b1e7a983dab9fc72cb9015c98e20a34fd3aed16 +size 266303 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y24q-Od40cuw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y24q-Od40cuw.flac new file mode 100644 index 0000000000000000000000000000000000000000..a6376986f659742d5bf3cd95562dfd3f7a445088 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y24q-Od40cuw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb85c2d7be50b7e274ef6780686c818943715214707cff8950990e0c2259bd58 +size 133656 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y26dNsDuIt9Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y26dNsDuIt9Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..8e8260d6c2bd539a88e99030d8f674a6d610cc1a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y26dNsDuIt9Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8face5499d7f79c0628ccd9bc38f151dc1134ee9b47201113f4bbec98e6ad006 +size 276116 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y28XfXlC7zkw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y28XfXlC7zkw.flac new file mode 100644 index 0000000000000000000000000000000000000000..ff282b3395f2ae18d8bec154119151fc673b58d1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y28XfXlC7zkw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e806bfee03706d3d97d34242bfb63d4b77b1f2878883aae61b8153eefaeabc0 +size 207898 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2AMWVZMp_Xs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2AMWVZMp_Xs.flac new file mode 100644 index 0000000000000000000000000000000000000000..7cd2a0bcad8c8d98b0b7e6b7c51be21779d1c2c1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2AMWVZMp_Xs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec0435f3f71695525ef967f2ef1e754088bd3b3256fd91c26d4121413b73f48a +size 240282 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2Awp9Nvsb2A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2Awp9Nvsb2A.flac new file mode 100644 index 0000000000000000000000000000000000000000..026896aee097b5fa57cdcb2241702c9b217e3bc1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2Awp9Nvsb2A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf78704a07b3fb13efbb3837ccbdb4c52912ce8816ef1dc2b76ea59ba83317a0 +size 210495 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2BZwHNWM0_w.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2BZwHNWM0_w.flac new file mode 100644 index 0000000000000000000000000000000000000000..42eed7d7ec68d767e91ca6c2d712f638860cf1d3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2BZwHNWM0_w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b1e64cee2f7f205dd06c3b39a8ed959198a472338ad478a26f921e0d924c0f0 +size 231359 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2CDH8aFgvu0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2CDH8aFgvu0.flac new file mode 100644 index 0000000000000000000000000000000000000000..6d0ae35f89c18014571d1601d24f1bc4514aead0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2CDH8aFgvu0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2bd955ff4604c3654df23fec64deaea17ca0b448f2b15b5b1c937c369a1f4ee +size 210274 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2E8eCKwj2M4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2E8eCKwj2M4.flac new file mode 100644 index 0000000000000000000000000000000000000000..63e52c00a1d39c01301a5e77379dea342d6fce86 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2E8eCKwj2M4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2184f81bba7beed71e53c6fccecbe5624cdb75905df052c1fbf09cdba29ba49e +size 214562 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2FZRNzNXEts.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2FZRNzNXEts.flac new file mode 100644 index 0000000000000000000000000000000000000000..b7896438d8a31d496aa398ecaeae71b5aa6c2056 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2FZRNzNXEts.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0f7443408f9efa50adaf9656724063121c76fa05e1d85880e6236aa8010756ac +size 219096 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2GBJOdvh9XM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2GBJOdvh9XM.flac new file mode 100644 index 0000000000000000000000000000000000000000..f168a1a0cd5a8307babb3c2d828cbed0b1bfef27 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2GBJOdvh9XM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af2519503a84f0a88a2459533eb373234422c9e5af95d9cfc7cfd1312f28a5c4 +size 224783 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2IQF6utML0I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2IQF6utML0I.flac new file mode 100644 index 0000000000000000000000000000000000000000..38b53cedc20eb57e21d3bf5304b469994510b5c9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2IQF6utML0I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d3efd53c32da12a1930456cbb195042f1ad27e7d52861e8d8e8c52299c5603d +size 205487 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2KR0C5ysO8o.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2KR0C5ysO8o.flac new file mode 100644 index 0000000000000000000000000000000000000000..5cfc8c78249150420c6ab42c013b8dfc6d5b3468 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2KR0C5ysO8o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b82bdfa2c08de808d920cfcad60c8c60553b3afbb9560e421b65c4917114abe8 +size 286386 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2My3EAryutA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2My3EAryutA.flac new file mode 100644 index 0000000000000000000000000000000000000000..e71ca72fb54a69d4d013dc3d18e1ef586e95236c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2My3EAryutA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae05493b212cef0487d60739b5e1529a4edeafbd6b4a2152028a4158589f448b +size 286372 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2N_9HjzCW40.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2N_9HjzCW40.flac new file mode 100644 index 0000000000000000000000000000000000000000..9c218e93d4c051be8a85ce9e7e30a4e4b76ee7ec --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2N_9HjzCW40.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:020dddcdfc749a328bf105da23c46faeae163329bf4e10663b5c1709bef3e490 +size 240659 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2PLQX17xFSE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2PLQX17xFSE.flac new file mode 100644 index 0000000000000000000000000000000000000000..c59cbd1d797cfd88c76f2ad4d4b6ca2b2d3e52c0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2PLQX17xFSE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87f7c2fb17f4778556f44bf664c4da9721a11b2efe90c0caedf17c7c87299619 +size 211427 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2PzTYy5GfqI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2PzTYy5GfqI.flac new file mode 100644 index 0000000000000000000000000000000000000000..abd1fa96bf590e73f21d17ad069dabbb90bb321d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2PzTYy5GfqI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:792e95473c6abf588248c43f8cebbda331551c1c6e87da7cc7c4d70c56127347 +size 172896 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2Q5NJPpcrds.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2Q5NJPpcrds.flac new file mode 100644 index 0000000000000000000000000000000000000000..5cda813a4c6f200d4a41109d7dbdcd2990586736 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2Q5NJPpcrds.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f96ab0ea498224eb4fc45b341842dccc0350aeb086aa11f88ec175633635bd3 +size 158325 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2QsWNp64ERU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2QsWNp64ERU.flac new file mode 100644 index 0000000000000000000000000000000000000000..4453d8bb13ce6652112b98f65bcdf2d0141d8bac --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2QsWNp64ERU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7637dbaad755bc45c9fc6f28976fce6c8e4fd5e8a48c39d02b32e48201ea8917 +size 247247 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2R9naYtL0tM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2R9naYtL0tM.flac new file mode 100644 index 0000000000000000000000000000000000000000..787385c3723e6b250dfebf358d45f177a28fc5a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2R9naYtL0tM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c155e10758d564f22df3634459cce063ab3dba814b8b93ca2c33d1dfd7b4d982 +size 212265 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2RTj-NQtrl0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2RTj-NQtrl0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c4d875ddbda5320509cd013d8d60bc3e2adf3066 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2RTj-NQtrl0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:caa4d31fffcf837b5aa97ed010b106e0e9b60075796168bd555a488fcd191b7f +size 200565 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2SELYAzRvN0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2SELYAzRvN0.flac new file mode 100644 index 0000000000000000000000000000000000000000..7ecd8965fdcb694c42e6d8ceb3d2a28c3bf9994a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2SELYAzRvN0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1e8e680516f3cbaf2d704a8ef8b7adaa1383ff19ebde39eecde30f66e57dac8 +size 230809 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2SfiNxoZzQY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2SfiNxoZzQY.flac new file mode 100644 index 0000000000000000000000000000000000000000..9491d127f9b431ee50a86033f0211d48def69c16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2SfiNxoZzQY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2bce836023a5928727a9e0b78fb374289aeedd3e1f33a72aa393f2572d039d35 +size 182757 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2VJSUZg-8Mk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2VJSUZg-8Mk.flac new file mode 100644 index 0000000000000000000000000000000000000000..55e97e3e7e3289b3a5a15b7f924932eecc9a82cb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2VJSUZg-8Mk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73ac48d4b582e785cce5ff465d28d35a0b727a5114ef66a8c392a63aa92e8f5c +size 237860 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2WQDTzY-v3Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2WQDTzY-v3Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..c2daf24074f43aca73537fed58921ace03e75d8c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2WQDTzY-v3Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1810cc8a4227489c3d6f6080cfca2957f99c43423b46d966d13634676631fec1 +size 201301 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2Xi_JpZkLs0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2Xi_JpZkLs0.flac new file mode 100644 index 0000000000000000000000000000000000000000..4127f2179dca993a307590b4fb28497355c63d32 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2Xi_JpZkLs0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3fda07bd065ef084f6ec597bbfa1558d6c25cb5436a3212ef5dbe1859770eef +size 177642 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2YzvStRrnkQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2YzvStRrnkQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..6df2fd84c392241212815a75b9611a35b7954cf3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2YzvStRrnkQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f792855a6f02b0b5fc8c4ef39dbcf50cb8e845ede3fe7fc158ce14886725d212 +size 203433 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2Zbl9J4vWns.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2Zbl9J4vWns.flac new file mode 100644 index 0000000000000000000000000000000000000000..526ff644dac362c54f283cf35f85beb62a207d1b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2Zbl9J4vWns.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0c3c4aa2828863b9ae592ce144430d75a5ac2b7708b4c6c07a285cdf04c0889 +size 249635 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2a0pqQomO8I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2a0pqQomO8I.flac new file mode 100644 index 0000000000000000000000000000000000000000..15e20daf60699d365ddb57fe540a7ac5a8fb6a5e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2a0pqQomO8I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d0c8ff5a8e3354bc8ebfc318d54fe4d7efd2e18ca81946b7389bf63872a06c0 +size 265025 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2b43Rkr-TO0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2b43Rkr-TO0.flac new file mode 100644 index 0000000000000000000000000000000000000000..6554b0a5bad665ea8c783cc6c6489f01511f3946 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2b43Rkr-TO0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc26834763e2222322ddc5995d574f9e734f5f40000ad67f9111bc19d8f61161 +size 176186 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2c3gzj2jIfk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2c3gzj2jIfk.flac new file mode 100644 index 0000000000000000000000000000000000000000..136cc6261e9889fd957fc84caa2c17056e480987 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2c3gzj2jIfk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13dbc1cbbf7c9b245109fac4ef3e034fbb8de198c7e9e3194ee3419c9517dcc6 +size 257975 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2eG8jc_R46E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2eG8jc_R46E.flac new file mode 100644 index 0000000000000000000000000000000000000000..bc8fe66fbe4bcc0bb1ac4220231dbd5bdadaee17 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2eG8jc_R46E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:740ca90af8e4c15561aca8d6bbc22efce2f2fe0528876b9e47ddc1b249010b28 +size 274809 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2er6dDbTYcM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2er6dDbTYcM.flac new file mode 100644 index 0000000000000000000000000000000000000000..ae503f4056a6e261f84fd45871c3fd0b9f7dd2eb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2er6dDbTYcM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:325d851949bf53df04f0f858705dc3a56c79d2d1499c43d30056895ef42e3c58 +size 217955 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2gMs40vtC6w.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2gMs40vtC6w.flac new file mode 100644 index 0000000000000000000000000000000000000000..68eae7c0eef74b7b2d10188e1b55a7220ae3ddba --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2gMs40vtC6w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10bab0a1da15169e44ae5202d3ca12428caa3dbe5351e631310a181c8553a780 +size 231035 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2h_F8niDeKA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2h_F8niDeKA.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5964b94f58507716ce624bad81fb244d42b966c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2h_F8niDeKA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c38bb13206122cd932ad9cd1cb07af76e70c2e00dde64bf2ee79217fe80aab31 +size 213036 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2lP3y7kW9L0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2lP3y7kW9L0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d483cc9785a8b496682e74c3a273af4890abc862 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2lP3y7kW9L0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6c9bfc241e817f66aa85e8bf795e324bda0720ead33e070601ac4903cd37842d +size 188043 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2o93RmaqiLw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2o93RmaqiLw.flac new file mode 100644 index 0000000000000000000000000000000000000000..1abb6073cf947a818092c7ed96fd22c905701050 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2o93RmaqiLw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e1a10e36fcd5105655fcaa384e5a790f1de13116484ede215eacfd27e5071b6 +size 232364 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2oOrBpVGJzc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2oOrBpVGJzc.flac new file mode 100644 index 0000000000000000000000000000000000000000..5d8c5e284655f6d06c7a8bbe70afa07520dbc7eb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2oOrBpVGJzc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97451914348e9f5b0b87483b122a9217ed3408acc999f43e97ca79ad8749f4a5 +size 234325 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2otvlKMxOQ4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2otvlKMxOQ4.flac new file mode 100644 index 0000000000000000000000000000000000000000..d6747c1f5c30f46255581335cd2468af2533bd1e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2otvlKMxOQ4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:994c8f2fa2ea93adf70036e07b953ceb23f6d87edd22e1b7ec0b4237a9a178cf +size 254919 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2r3RbvqKYEQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2r3RbvqKYEQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..904b0cb5f09d2732c8cd79fa3a830063df57bbe4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2r3RbvqKYEQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42f23557df4d818fd743c111d3bfccbf895d2c15b87fbc98c62ab2da25a16ae2 +size 177037 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2sadF79hT7I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2sadF79hT7I.flac new file mode 100644 index 0000000000000000000000000000000000000000..10eb25c2817e07855aacb628907ea7440716cece --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2sadF79hT7I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e8704564867e39d7e06620af5f1f2a949781ece29342ee3df6b8dbf2ead247c +size 238110 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2swvx975qSY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2swvx975qSY.flac new file mode 100644 index 0000000000000000000000000000000000000000..50458e6590e4437393bf93fd131df29055873188 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2swvx975qSY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffef85eac04a1e60be7eeff715f82005507d2c7455f4e3f4053fa3c83e370646 +size 150273 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2uCe5wv_iQQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2uCe5wv_iQQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..779c18ad1be9c4bfc9238b02968e4a866c0319ca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2uCe5wv_iQQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9034c5fffb92247f4f3aa8b30231068aa9ac1d545801e21d07f6cf4783754108 +size 154985 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2uOCHpAzwXI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2uOCHpAzwXI.flac new file mode 100644 index 0000000000000000000000000000000000000000..ffeb29d637f4849c378481b7ee4766b0083578d1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2uOCHpAzwXI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d55d51739e48ada87f638cb02e531389f7f1a418eb99bfc2d87cdf1c1f01f12f +size 234065 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2ugsMuS9EXE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2ugsMuS9EXE.flac new file mode 100644 index 0000000000000000000000000000000000000000..50ee88a4f7c89f78d2827d0a62014fd5347b4d01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2ugsMuS9EXE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34f94790ed15c0b888973fa5503cdd2e0f40f30fc7431b3a9e913c43189901ff +size 175725 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2vBbpFbRHmk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2vBbpFbRHmk.flac new file mode 100644 index 0000000000000000000000000000000000000000..a26c1c8385e319bd8c6a6f2614fb6cd46a7d8be0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2vBbpFbRHmk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d27d65279288ff8506a2b4ae623d6d7409f16d3823d894bf77a09d75db01cab +size 156218 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2wKP1A8On1E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2wKP1A8On1E.flac new file mode 100644 index 0000000000000000000000000000000000000000..e484e09ae99e560fc42a84f0de5dae1537d344f2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2wKP1A8On1E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f245db1c6e4d38f0be47e163375a356a79bcd9b9b3336b51f34f1c55b3ea282 +size 231486 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2wNAe_DNXI8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2wNAe_DNXI8.flac new file mode 100644 index 0000000000000000000000000000000000000000..db6fc185ce89e7a90935db83df7f4f4b84f4a97a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2wNAe_DNXI8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9749d08bfc3c90fd6d818d99fdb4164ed02ce75bbfcdd1a94d9e2e3cbb741ed3 +size 220466 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2wz_XyJ4Zxk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2wz_XyJ4Zxk.flac new file mode 100644 index 0000000000000000000000000000000000000000..836d3c970876775116a4459b5293eba9cea44a7e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2wz_XyJ4Zxk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:926df5f7a962709c330f354322aab70aa1fb6405a10cc7b47236aaeef8b1b6fc +size 150511 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2xg_8rf9J50.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2xg_8rf9J50.flac new file mode 100644 index 0000000000000000000000000000000000000000..94d5c99c8246e30d0b3441a24343f00d51ee073d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2xg_8rf9J50.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c6994fb6b26a5c309b7870a6d6db03167609414a4420afd7b1a3f8a0e0fde9e +size 202609 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2z2ekcBoxRs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2z2ekcBoxRs.flac new file mode 100644 index 0000000000000000000000000000000000000000..ff9b56f07fd6c3519ff31cc062a5e133b260f9f8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2z2ekcBoxRs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:507d1322a8d42b3dbc9a5ff26a8c828e7425891ce4718ed2eb675fd808c7c1a7 +size 214269 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y2zGT7SeSHXo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y2zGT7SeSHXo.flac new file mode 100644 index 0000000000000000000000000000000000000000..cc10efbb371d5014ed8dbe17cf20b5d3bf3e40a3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y2zGT7SeSHXo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa716ce295da41635b48ee0c7d8ec5e5e8698b4141468104794b8f0b5682dc44 +size 175676 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3-9kjeKeew8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3-9kjeKeew8.flac new file mode 100644 index 0000000000000000000000000000000000000000..eb44bb7b033cbe34adc3e02811586f7ad079866c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3-9kjeKeew8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43cd54af52df016740d2f37e48be32d6756f3232ef82ddb4e9dbd47a0d82b752 +size 175110 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y30D1aU785Gg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y30D1aU785Gg.flac new file mode 100644 index 0000000000000000000000000000000000000000..17e8eaeb3f3f89177d93adda8686f905c64c1263 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y30D1aU785Gg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57b350ab8bec9f967dd3134d16badc6f1449983480fd5c1d28bc075d71005812 +size 142529 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y30Sm6NPl_rI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y30Sm6NPl_rI.flac new file mode 100644 index 0000000000000000000000000000000000000000..3e4ebf5af0bd14d5104e8a3f0e853ee45cd8e4d4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y30Sm6NPl_rI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bb1c6ded547bba0698ea4bbdc6443e3ff90889a2065f0dd06c991d6be4343c9 +size 229651 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y31oUjXfb1Gk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y31oUjXfb1Gk.flac new file mode 100644 index 0000000000000000000000000000000000000000..3cf248d81f94341d07888f181156108d7ec9c8e0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y31oUjXfb1Gk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:780063e56520b36bffd3af0f134aa37769e96cd057b07c4d51f71d693fad48a9 +size 165772 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y32T8xTpvTFI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y32T8xTpvTFI.flac new file mode 100644 index 0000000000000000000000000000000000000000..4610e6a363826db82da0b3a8d3201576f5450089 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y32T8xTpvTFI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae3e55e65a06f2268991a6a8840c4bdf28d824f6402684e4d460ed0117c56147 +size 223438 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y32ptEe--IM0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y32ptEe--IM0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c9d3a6fea56d1f2fbbbcc65b915f84c5b69a95c1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y32ptEe--IM0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:370afbb0e5c372804535ffce63b7060ff714bbbf3998e80a739f41c1238fbb6d +size 184923 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y32vu_dTOVZM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y32vu_dTOVZM.flac new file mode 100644 index 0000000000000000000000000000000000000000..a67e507745a1104c85c6b0d396a6570505ac25ae --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y32vu_dTOVZM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3a24fdbe723472140700f20bbced9d077af3d412d366d774b3fff8f5f8fed2d +size 196634 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y34In3a1ihmA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y34In3a1ihmA.flac new file mode 100644 index 0000000000000000000000000000000000000000..fb4f06a78dc4c0b51c88782805ecc21b617b63c3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y34In3a1ihmA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7afb2452bb2e327ee02f0e63300d9e40e7d7d72d4b7625acbbee5a6ad09fa136 +size 208273 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y350OCezayrk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y350OCezayrk.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7977b48483b1bc43638b1cf9edb66963746dfd0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y350OCezayrk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb32f9563ab7fd158ca651e3925efb2eb86cf9e64e07da3d13d56181c0b8e543 +size 272362 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y35GZhbIN5lg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y35GZhbIN5lg.flac new file mode 100644 index 0000000000000000000000000000000000000000..78cd0b44c748491972ddc550fa45ef8309b97ad8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y35GZhbIN5lg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a2f5135b7ecc37b8b0dae3b5819b92b52b58d32c614f5b5d01f7079b5da2807 +size 205310 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y35t3MVCE160.flac b/Foundation/Audio_Grounding_AudioGrounding/Y35t3MVCE160.flac new file mode 100644 index 0000000000000000000000000000000000000000..fb052e7b7512f608324128354c0ebaf61e27ff80 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y35t3MVCE160.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5510278a66909b11faf1565ca7850bc74010856772465baad2fbe5eddaed13a2 +size 239585 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y38rpGMJA_HI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y38rpGMJA_HI.flac new file mode 100644 index 0000000000000000000000000000000000000000..4ad8b684246c535441609ccc9b76b84e983df4d1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y38rpGMJA_HI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bee8f88f564eef53b819e7c4649faa94a8d250775e3e20c166605237a53659b +size 108015 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y39EiNqi8feA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y39EiNqi8feA.flac new file mode 100644 index 0000000000000000000000000000000000000000..9bb2bd1a3bb1db48a58b5e06acd66015e2d4709c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y39EiNqi8feA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:260a96ff72252bad3229120d1cc393d1abec7b4e6e4a7c7094e22c0a7fbf4707 +size 243270 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3Bk--DQpmUs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3Bk--DQpmUs.flac new file mode 100644 index 0000000000000000000000000000000000000000..e1bd5312d2d858190a7e955b97082d933b6927d3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3Bk--DQpmUs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d904ebdb0233d24c4ef0c4e8a96d9d2ae82bec32b7caa0fe5c5aa1a1705a1690 +size 232659 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3Bl7Ko2xtmE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3Bl7Ko2xtmE.flac new file mode 100644 index 0000000000000000000000000000000000000000..44ffc66a6b599575310a7097459218d8ba9c763c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3Bl7Ko2xtmE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:340484e894c590689fd979f7a1a9aaf2efcc57dd3d807e7ad74abdddc3d42a29 +size 138525 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3C9e_UT9ASs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3C9e_UT9ASs.flac new file mode 100644 index 0000000000000000000000000000000000000000..13e7d39361a59feb7f2474d66853cb4a303742ad --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3C9e_UT9ASs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fc4490168ff2ddb0d4bc3e9ab8b44c3adc1d4071b53a6ab6c34bb5388cb7f62 +size 176025 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3DMi_CmlYJ4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3DMi_CmlYJ4.flac new file mode 100644 index 0000000000000000000000000000000000000000..c7dabe3184b781aef24dd595ebe794ad8aa4bc75 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3DMi_CmlYJ4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83c154cc2827c8edf63a0e946ebfffa2402f5b7cbcfb2954e3504ef0758c4abf +size 234911 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3G2GK69lD4U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3G2GK69lD4U.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5f28c4acfc2d662993a745cc238d0aca51d20db --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3G2GK69lD4U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11840a6e183830b6867e8b9b33cc468f68a76871a7109c7454ab8b234b03c380 +size 230506 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3Glt9M8l_OA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3Glt9M8l_OA.flac new file mode 100644 index 0000000000000000000000000000000000000000..bccc79c2f75ef7b637c0d3d8bdc21ccf2a41ddca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3Glt9M8l_OA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:615f1107e7fac7cf62c5125fb8aa8b09f24eb793fd3c27196a81a7c0b9174383 +size 140944 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3IPOYt95c_Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3IPOYt95c_Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..7fc4a39b7182e4983fefc14f967428ec41e49edb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3IPOYt95c_Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab85d833ae57439309e69a1c5a661791e03ca755b68d9c40209115bf01fddebd +size 212952 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3JVI9QiahhE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3JVI9QiahhE.flac new file mode 100644 index 0000000000000000000000000000000000000000..9d76618243ee0eb626508c61309c45af7bb33be6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3JVI9QiahhE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e03418647ec9e8d18c43286ac8ee76ba0dc2ab40343d4d47605f2c348be761b9 +size 224944 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3KF0xSdYWeE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3KF0xSdYWeE.flac new file mode 100644 index 0000000000000000000000000000000000000000..0d5b1081a05616eda01a9afd4f8a7aa2756aa1e0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3KF0xSdYWeE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb707cc66672abe7bad466bb43cb46a367dc0290d36573e8fb67b53d4d2f7e41 +size 183579 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3KdlTPIqOpk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3KdlTPIqOpk.flac new file mode 100644 index 0000000000000000000000000000000000000000..18001e0dde15ae2a73dca95b107fe3b04891aa27 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3KdlTPIqOpk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff342432ad6dcf93305dcca4eafd409b88e14e3bb2dd825df11c8d0ffac431c8 +size 223292 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3Lcxs4pPY9I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3Lcxs4pPY9I.flac new file mode 100644 index 0000000000000000000000000000000000000000..31fb74904294ab83e7a2c6577c60d6090b114fa7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3Lcxs4pPY9I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8cd5673a42460c71e8cbc49293b4f4eb6c17bc9154c307f7f683002b808e44b6 +size 143075 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3LeU8LHBvnU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3LeU8LHBvnU.flac new file mode 100644 index 0000000000000000000000000000000000000000..c6a96cea40eccd3f6174d0e451659c5250f068c7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3LeU8LHBvnU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:460b73721b5d2159a66f42496c96544c6b63abb273cebba17942ebd7da65fe9b +size 222779 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3NNCGNtAlVE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3NNCGNtAlVE.flac new file mode 100644 index 0000000000000000000000000000000000000000..563a2be1a0563009422ad78cc575f7f3ad7dba9f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3NNCGNtAlVE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1333f00fd09929e7bdf92b442823d31725764ac6610e8e73cb8790ad4432800 +size 283778 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3NPidVvhTnI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3NPidVvhTnI.flac new file mode 100644 index 0000000000000000000000000000000000000000..dff1bd50690e7daf6f57803a7c68f30913ccee72 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3NPidVvhTnI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0448d5994c2609c210a2aa6a5c11aeb959f75884e6f3947ecace095a113542d +size 256213 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3OFi-JWcGK8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3OFi-JWcGK8.flac new file mode 100644 index 0000000000000000000000000000000000000000..fcd07e59cd807bc39aebab1f0c409cf0a0d67562 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3OFi-JWcGK8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63c4cf0bcdf25e1092f945d95ae8b615730a3490d961f03e629ccd2512eeacd0 +size 197652 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3OeT2w-6B94.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3OeT2w-6B94.flac new file mode 100644 index 0000000000000000000000000000000000000000..b6b2ff7f15e391b5204d59a614bd589a4b769733 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3OeT2w-6B94.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c84db5803b2e218d6540668bd3372f762cb00deb8017eba04d5040970184e38 +size 186577 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3P1EgAdpL98.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3P1EgAdpL98.flac new file mode 100644 index 0000000000000000000000000000000000000000..de498e62083896c5152b8b8d1d0068482201eb6d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3P1EgAdpL98.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33bd72160eb48254a1b9aa7ec193d7b8cdf83c151a6ccac8df555ea9db71c541 +size 136259 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3PPL7SS2C1Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3PPL7SS2C1Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..373c5c0b2bdc805bf1926d24cc74af8b9783e66d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3PPL7SS2C1Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8f87c0b9490d11c5c56471b49b0f2c90b0d09351f74103f79313195fd674f48 +size 258003 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3PVSh7dth8Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3PVSh7dth8Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..37cefc7adc3824a66a2b016273623172bac3466a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3PVSh7dth8Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e666a9746d0eedb77056957a55f0e4f55f765d690ac3abafaa6e8fd3690d2293 +size 171303 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3RPKGqAHRiU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3RPKGqAHRiU.flac new file mode 100644 index 0000000000000000000000000000000000000000..5041e3931c78799f3511f216044304cd4e192552 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3RPKGqAHRiU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ff5efa742fdd8b0abcf3969b2bdf6da42171498925a87b881df623a597d17276 +size 214222 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3RZySuDzocQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3RZySuDzocQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..5fd80b8857493f1b8d187dd2494cbe84d81a16ec --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3RZySuDzocQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a9d9e8a88f5d83da179470cab16f989d975a4814e8b5da0fd4dd3a571ee56018 +size 244475 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3SFet_JSgfY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3SFet_JSgfY.flac new file mode 100644 index 0000000000000000000000000000000000000000..5e10f946a5e8cf689228143c85bed6323fa04337 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3SFet_JSgfY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3d04fe001e0e3bb191f7d3a1adb9337390ef283cde27e97bec19a1de8c2264a3 +size 188846 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3TZTVzCsPrc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3TZTVzCsPrc.flac new file mode 100644 index 0000000000000000000000000000000000000000..b50e9ed67eafa5bca3f8a15667cc9d2062790d5b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3TZTVzCsPrc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:af9a2faaa72bb605fb092ccb99db6165adf69deff5c12bba172b31af1660d435 +size 258463 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3U1ZravTj_Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3U1ZravTj_Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..a1b8ec13f19d7646d045a0b36df22d1ea71c84d9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3U1ZravTj_Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cf1949631ad6029c4daab2ab284f0a5d19c0be2c9362a4b8ef1a742af03e1b0 +size 183507 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3UjD8ncozAA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3UjD8ncozAA.flac new file mode 100644 index 0000000000000000000000000000000000000000..718d6b963d1b613010c9df488f80c6ba03637548 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3UjD8ncozAA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a542681ec9a6c9dcd99576c9faa65e896e646c269327f371dda1c8901e4cebf4 +size 156855 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3VfzEumJgcQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3VfzEumJgcQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..67d3ee96977ce25bcd59bb89653a8c264f32221d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3VfzEumJgcQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33d3fabae2beb858a086e4d4d0cdf8173d8b8df770ec1b1fbf0ef6013b18aea8 +size 222807 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3ZAtv4S5o5Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3ZAtv4S5o5Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..db957d0d46650f386d2d2aa1a66d4b0722cf0ed9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3ZAtv4S5o5Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48a4d254fc2a862d5a8a18ba1fc1e84b1eb4fb4d1be0c3a6199b4c035eb2f201 +size 180876 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3a3cnDXnb1k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3a3cnDXnb1k.flac new file mode 100644 index 0000000000000000000000000000000000000000..b486f89cfd594623b7618ca2a1f01cd78bbd7360 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3a3cnDXnb1k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efa0f834e00da9d5e519de84a850b8622a77470be747c6486fd57e2ccfb19c24 +size 224657 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3dFdo47pK1Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3dFdo47pK1Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..65c8afe8bc669a25709d290598ed9db2a39af267 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3dFdo47pK1Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b6be311d121e94e4be6be38bd4b35440d14d003e6c1c8fe60efa3f3d54cfca9 +size 107518 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3ek_TD_ZAWA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3ek_TD_ZAWA.flac new file mode 100644 index 0000000000000000000000000000000000000000..ce894be4f5ea3a381df262aadac7bc3f91c3629b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3ek_TD_ZAWA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65c2a22ea6f994d16a16222bed3404809916fec3fb12e04c8d0393d92a2a0da6 +size 242943 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3fQyxJQXF3g.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3fQyxJQXF3g.flac new file mode 100644 index 0000000000000000000000000000000000000000..d3966e3a8372377b65e8cb7208747e31ad045291 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3fQyxJQXF3g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9857f732b0eb6ec79a3fb16aa11e6c07a1604b135a803f918b43d36942abf85e +size 200491 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3iKaVJsj6co.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3iKaVJsj6co.flac new file mode 100644 index 0000000000000000000000000000000000000000..620bab636f5611a3d748635dc8594460ebf40570 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3iKaVJsj6co.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab36047c6d7ac8475c6744e57eeec6d4378f9b9106688244f12aabe8351ddd18 +size 236696 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3kNCkzSBrfM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3kNCkzSBrfM.flac new file mode 100644 index 0000000000000000000000000000000000000000..9b28a2e0216a420cb734b44e307e91104c7eb374 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3kNCkzSBrfM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51bdf068bb0e0bb1de2a959aeccada35a64666ef61ec4d931e28cb66a6c16602 +size 279729 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3kbHcjynq7k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3kbHcjynq7k.flac new file mode 100644 index 0000000000000000000000000000000000000000..711ac8906c49f56769e5ddc07282ab882103c649 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3kbHcjynq7k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5731b95ed23f10f82f5e2852f690c342b3912d46e6d72fed3c7acdf93f19fa33 +size 258183 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3leh-_Xki-0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3leh-_Xki-0.flac new file mode 100644 index 0000000000000000000000000000000000000000..8d18587e4076b75bc3940952e612c375b283fc16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3leh-_Xki-0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:946955a2051f94171a75a9b8082b99fa5c7fa84e8111d664afeb5e5aa5d8460a +size 207031 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3n4G4MbYRFE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3n4G4MbYRFE.flac new file mode 100644 index 0000000000000000000000000000000000000000..44ee452d6ddb500caa614fe33b8bfa927207b832 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3n4G4MbYRFE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6ef529b5a6f9d6f6b7aa6109f9216ecce013282a82a6947880573b252140bb4 +size 235744 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3ndid3jni7M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3ndid3jni7M.flac new file mode 100644 index 0000000000000000000000000000000000000000..700a7fc419019035ebcb0d5bbc8beed50aca9d16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3ndid3jni7M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dcaf2b5a0f5cdc838b9108ae7af0a1919c014fe52a1585913a67e62976ef119e +size 222478 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3ngXHZYzxww.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3ngXHZYzxww.flac new file mode 100644 index 0000000000000000000000000000000000000000..5dd84e2ca5619e596cbb833dd1720f87725fc84e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3ngXHZYzxww.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b89dc1bdb06b7443b51662ffdcdaae692ba693edcb4966fbdf832dd5357f66c5 +size 203837 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3obLPXCKYpM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3obLPXCKYpM.flac new file mode 100644 index 0000000000000000000000000000000000000000..c6279b64a3dd5900ec464375362560e1eb23ab32 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3obLPXCKYpM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:302ea4b720a6d9b36bb2d0f7a4bbde50cfd155c2a060612a5493da357faee9df +size 256997 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3pLkwWZxvYo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3pLkwWZxvYo.flac new file mode 100644 index 0000000000000000000000000000000000000000..a3b50878aa8ac64bcaed2e7406e41d6f37fa5a50 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3pLkwWZxvYo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00da679e28cf075ac123d5f99226c28c8025dd6da872a1a9eb3f0f6b1b00827d +size 278526 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3qnpOqvnKTo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3qnpOqvnKTo.flac new file mode 100644 index 0000000000000000000000000000000000000000..49fde7967aef50c6e41e79415e81e05456362c1d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3qnpOqvnKTo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0bba1b6107eedbbc742852bc70a973c1b1b607564ce86102a80a27d38d5b1f4c +size 262649 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3rAJOlCARSg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3rAJOlCARSg.flac new file mode 100644 index 0000000000000000000000000000000000000000..7c242ff002f59db0257c041cb30a8a6d2b659c5b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3rAJOlCARSg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:215cfe69dc947132e0275cb2a41af15f8181add9e2a5e89b404c24ed33ebbc4e +size 236802 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3rUTwDYoPDI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3rUTwDYoPDI.flac new file mode 100644 index 0000000000000000000000000000000000000000..dfaa7c0b05cfb7fb89b09d3f01d81a36eb74e307 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3rUTwDYoPDI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e11a479fd76a88c2571ba417f0565efa047d88d162b1ad1ed15a86a11eeb131c +size 200278 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3rfkzWJqq7U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3rfkzWJqq7U.flac new file mode 100644 index 0000000000000000000000000000000000000000..4e3a8dda1c43caa74f979f38f622a7e92a0c337c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3rfkzWJqq7U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bae14bc489bbea0d61d7e04c6bc58a79b20670ebb36103730627e19c74584427 +size 232196 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3uOXUGT0kxY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3uOXUGT0kxY.flac new file mode 100644 index 0000000000000000000000000000000000000000..cb6cfbbe312021711f7e0f78e9bcdf4f999da8b8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3uOXUGT0kxY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd74c8f09af25ad5d069675c923916c56413d22c52c18d48c19cc386124a7765 +size 254911 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3uSEbfnfPUw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3uSEbfnfPUw.flac new file mode 100644 index 0000000000000000000000000000000000000000..a02195a2e8311a1910fc7c531529be2602e5cd33 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3uSEbfnfPUw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58b74bedf367b7d8785372e346502f448320997969dcdf033c21b6e7c0915b3e +size 180717 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3ujCBi8VbLc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3ujCBi8VbLc.flac new file mode 100644 index 0000000000000000000000000000000000000000..f503d685d34e39836c2bb52b867cde9737e62130 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3ujCBi8VbLc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c0ce117c34475b783ee6e15b038d9f19ecb23c2161925133a51bd3076679477 +size 127029 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3wVoCVgkqDE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3wVoCVgkqDE.flac new file mode 100644 index 0000000000000000000000000000000000000000..1e3d83b273ab78fd2d53d0e4f758b9d99133e269 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3wVoCVgkqDE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51bd81e5fd266dffaa871938b936cf8072c6772c58827ba3dd3617c4e936359f +size 137376 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3wabvpoDwKU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3wabvpoDwKU.flac new file mode 100644 index 0000000000000000000000000000000000000000..bcde63f1efa3dd7a3f8a74195114d6d373a444c5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3wabvpoDwKU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f17eda2dc7c35e978b0b766fd03b59b26e32efd6865f31a00359cd8dd544f501 +size 141137 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3xLintYp7PY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3xLintYp7PY.flac new file mode 100644 index 0000000000000000000000000000000000000000..640e1e4bf4aec81d8561ff5b7cea1b693c9335fe --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3xLintYp7PY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5fdd65546b5e6bda9955f4208153c930ecfd0f4c9c061081f0db1afe7682b5fd +size 112277 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y3xepWaHuMN0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y3xepWaHuMN0.flac new file mode 100644 index 0000000000000000000000000000000000000000..9b87bb8ff192bd3408e6d801b63f3a9c5f489e3f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y3xepWaHuMN0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58ae1dbcf1c1c5f15a3957c655a1c311674151a57710172696e44681311ca22f +size 200931 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y40KPvO4y9M8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y40KPvO4y9M8.flac new file mode 100644 index 0000000000000000000000000000000000000000..cb754634a3ecfd35a33c043701160254224dd920 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y40KPvO4y9M8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae16c68d73e4bcf8994a8fa7ba0f237a6d183dc1173c5793c4d77b9b929320c0 +size 252867 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y40cuHrYfaqA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y40cuHrYfaqA.flac new file mode 100644 index 0000000000000000000000000000000000000000..6256879175286cb2fc819c8da727581f3fe8c894 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y40cuHrYfaqA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec0176a602222b6d36f298b993b5e368d2f41ead3d2a6e33acb22991feed682a +size 67969 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y41LCAS35xP8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y41LCAS35xP8.flac new file mode 100644 index 0000000000000000000000000000000000000000..f684974d16773a1ba67e02a5de82156d95b82b44 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y41LCAS35xP8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa30d9b61e4fe49425f2070deb96d14b4902d52a68d275f3dfdc006171fb69fd +size 212264 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y41_uGt-6cHI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y41_uGt-6cHI.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7577495b80cb66947b29b3b1925f56b26649074 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y41_uGt-6cHI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00f562824573656c3ddba97e67870ae412354973bcbfc85791b488c413a8c2c5 +size 245368 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y45KEHJraS_I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y45KEHJraS_I.flac new file mode 100644 index 0000000000000000000000000000000000000000..300e8f243cb4e607a2a4bfabcc916b41047236da --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y45KEHJraS_I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df31f1de33e0ec6a2a42cf794b5013483239ec701f05217c92c2d58a28c35096 +size 216353 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y47tADKZxUnM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y47tADKZxUnM.flac new file mode 100644 index 0000000000000000000000000000000000000000..05888ae1590949a1ff66dc1913fff5c96a317a41 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y47tADKZxUnM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3a941fff8d1ccb721926ad99376a319729dd035cd2080cc43f0741f17a7f607 +size 221728 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y487WZLlab7s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y487WZLlab7s.flac new file mode 100644 index 0000000000000000000000000000000000000000..5b741cab75438cac93a11136dfc56c93334d54b5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y487WZLlab7s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f37de792e4ae6048c160ae3265dbfc3a8f6d3cc9cfdd25e700ffafb22f0105b4 +size 188380 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4AvF_quFNSg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4AvF_quFNSg.flac new file mode 100644 index 0000000000000000000000000000000000000000..90294ca6744cf7b329aeda6a8e65e3c9e8dd91ee --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4AvF_quFNSg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4447d5a4d2e5dd11b0128f2ac969d2c8d80dbe8baf2627396f3a6632677befe +size 262994 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4CZ6vWHPVr0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4CZ6vWHPVr0.flac new file mode 100644 index 0000000000000000000000000000000000000000..3ee655d019d2285761b72156a95ec2b7e85118b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4CZ6vWHPVr0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a128fb053288b9db514b51711725795b94d8eef18502b9ef4350361200cb143 +size 279170 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4GFeZgTcgRg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4GFeZgTcgRg.flac new file mode 100644 index 0000000000000000000000000000000000000000..2447d6f36fd47eac37209ecdf97a948f107cf0aa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4GFeZgTcgRg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b7b2ad29cd488eb42c6e1800d69034af64c91c25e8a68eb581d159f81e94ef82 +size 186123 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4GalkerBVs8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4GalkerBVs8.flac new file mode 100644 index 0000000000000000000000000000000000000000..ac103ba2ec8fefb7889234ffa1e9847ee3826166 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4GalkerBVs8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:297d046b79a8bee1bca634afc07cd1e0b9e669ea75d3e9eeeeccea4891c5526b +size 148942 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4GfN0gZQvXs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4GfN0gZQvXs.flac new file mode 100644 index 0000000000000000000000000000000000000000..3f391ebcedc26ac64cc23271efce00729e904a3e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4GfN0gZQvXs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05f3c07385bf9ddc78db3546f6909b577eeb48b061b65888375e9e730ec3e102 +size 217559 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4GzG-sDAwmI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4GzG-sDAwmI.flac new file mode 100644 index 0000000000000000000000000000000000000000..f76f15f4c7912854446da63ebeb459055382682b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4GzG-sDAwmI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd74420ad8c63e40f29c506f8fcc765dd90f8eb5b41c5b7c91d2582d891e0004 +size 288115 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4Ka46DJjDZk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4Ka46DJjDZk.flac new file mode 100644 index 0000000000000000000000000000000000000000..8b4bc8cb03af7062869b604e0fae7e962ad0d6d6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4Ka46DJjDZk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf3a12665d9d75d4369880f23090284dc8daddb579d2ecc769d2db3b413267c8 +size 210893 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4LV2zS3hJ_I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4LV2zS3hJ_I.flac new file mode 100644 index 0000000000000000000000000000000000000000..dd2db7ee586e3f5a34015b513aec0841898db1b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4LV2zS3hJ_I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3b7e2d05f00be231845325b4c0720229478e948a20a45f29ced58878f14c4190 +size 247314 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4N01YFev3lo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4N01YFev3lo.flac new file mode 100644 index 0000000000000000000000000000000000000000..8ba71d8a2c3ad09792c1d1b8e6716923c9e4a113 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4N01YFev3lo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83f1e3adfe7c83cb987f826204c93f162c9e6242e1e3ebb99ff07db1337e190d +size 185979 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4NhrtAEUsnc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4NhrtAEUsnc.flac new file mode 100644 index 0000000000000000000000000000000000000000..0fe569c81b36ed623b9a2eefc56686c4d245d93d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4NhrtAEUsnc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c6e82cb437b84c5ed8dd034cd1ab61f6934db0f265a0ac1959d7360870bbb4b +size 291554 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4OIHJwcKSWI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4OIHJwcKSWI.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7fc58c5d0491e43b25b9d8f6b369f8c20ab6c0d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4OIHJwcKSWI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f57128a59ef08c254875cc4178fb38234e5196900136cb91855056907b420591 +size 257152 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4P7p_EIdIeY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4P7p_EIdIeY.flac new file mode 100644 index 0000000000000000000000000000000000000000..b31c9d5de268ea622c0c02f4c9fd3a177250993f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4P7p_EIdIeY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2af6785c53c3b73ea018debcd1079eee7c36779b08d802029171524568a06f9 +size 183435 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4PcRN6DC7co.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4PcRN6DC7co.flac new file mode 100644 index 0000000000000000000000000000000000000000..6d592dea5f9ca97d1f234d280ff442aa7654391a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4PcRN6DC7co.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01fc4e23a178af2ca9aa8d62b09a9a37ed5c488ebec7c14e831a4c1fea314364 +size 229235 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4QEON5KWC1o.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4QEON5KWC1o.flac new file mode 100644 index 0000000000000000000000000000000000000000..35d57f57926a5c83047fbe2be01ee035626642a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4QEON5KWC1o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3e1c786033c31d716ec1b4cf7430f88e9f4a707a3910abcceca53e5789d80e5 +size 202869 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4QHVLFkpiAI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4QHVLFkpiAI.flac new file mode 100644 index 0000000000000000000000000000000000000000..8c961c0f94efdb994720adab276fc136a5eb4fdb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4QHVLFkpiAI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e1a97286db8d8b6deab47d838fe52458995765ef95ce1c78d6a382c65848991 +size 242613 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4RU8_lp0Z2E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4RU8_lp0Z2E.flac new file mode 100644 index 0000000000000000000000000000000000000000..129514ce932c5ce887098cc5e453fd2fbde372d7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4RU8_lp0Z2E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23f21776219171b4ac8fc048401d6d9f2b1ca70cc649bd6b641c3fcee6dea3f4 +size 235903 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4SGVC_RNijM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4SGVC_RNijM.flac new file mode 100644 index 0000000000000000000000000000000000000000..bf465b7344757ed46928bed112a067160ffed393 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4SGVC_RNijM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1a770023b0c724ab11ec94dd105719726a6a3bad515619b4369a42e8eaac08f +size 264582 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4SbXNmgnyio.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4SbXNmgnyio.flac new file mode 100644 index 0000000000000000000000000000000000000000..42aa3a44d28cdbdc8a53c5888841eeedf2e22fb3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4SbXNmgnyio.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2498228b1260c07d8c94d0dd419adf1d5906c35acf2173c49019d7ecfac3b03 +size 262487 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4UTdag5Lp0o.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4UTdag5Lp0o.flac new file mode 100644 index 0000000000000000000000000000000000000000..37da8d99bb4e3c6e8478e23d124160ae57b96a20 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4UTdag5Lp0o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:857c4ce95e3948d633d9f5e1e4875d6910db880d416e950824c84634a5c26830 +size 232583 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4V3TFiHFqDc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4V3TFiHFqDc.flac new file mode 100644 index 0000000000000000000000000000000000000000..7551076f83559abecae0434599cf729fd06b8166 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4V3TFiHFqDc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:536112e064cdb34cd3a68cbfa90926f77ed9ee39e7ff8bd9e5a245a04ef561ef +size 230681 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4YUp45FEi5U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4YUp45FEi5U.flac new file mode 100644 index 0000000000000000000000000000000000000000..e881aa63ed78a1428c29f7c5b392365452fee976 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4YUp45FEi5U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca423dbde806164569337c92f5e9debe258cd21b8f56e9b589a8628cc44b7fd9 +size 203885 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4Zcfo6EtDww.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4Zcfo6EtDww.flac new file mode 100644 index 0000000000000000000000000000000000000000..55fa8140b2de94080e14801e40190d42e6e260c7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4Zcfo6EtDww.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54e603bc117e55ad011247c7cee4f24c41d7574591f4b5df722354d523d2426a +size 156686 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4ZdX1Hl6mhs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4ZdX1Hl6mhs.flac new file mode 100644 index 0000000000000000000000000000000000000000..9863c131aeff677827a6c15d63e8421ea0f6fcba --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4ZdX1Hl6mhs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abab19571fbb8d493a66575fc94e03ff6766f0bbd26eb89f0e2a8fe82e5a5811 +size 181880 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4_bKiFo_znw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4_bKiFo_znw.flac new file mode 100644 index 0000000000000000000000000000000000000000..7260e0f537f8a4e6a0d39da6a141e8d0d52af857 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4_bKiFo_znw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d9e2368a6702b788d29728885a3f81d75f081f1290321b451a6ad4289db7613f +size 178526 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4bp-AKTPBYE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4bp-AKTPBYE.flac new file mode 100644 index 0000000000000000000000000000000000000000..861e0e9ca1a05548ea11d85553d1d9aa527f41b5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4bp-AKTPBYE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4879c0d1435c69d98eec5e1ed3352750ce928de1e749bdd45e74ea549c3c746f +size 200938 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4dgvqmpMORQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4dgvqmpMORQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5fa3834a6b0565a77f9390b6db5ff743f9b9091 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4dgvqmpMORQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa304b11931a7b3e38345f8e1a0745170e1a4dcfa81b6b90d57cf3b7b67e128c +size 239917 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4iwHOSTWd1s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4iwHOSTWd1s.flac new file mode 100644 index 0000000000000000000000000000000000000000..f4a48f5063605afba6d441d1dc6cd366739230e5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4iwHOSTWd1s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0042ea140a84c8ea9f74f56647e6ee7cb11b3d9061b65928fad1f56568b4f085 +size 254709 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4kawUPzXAlM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4kawUPzXAlM.flac new file mode 100644 index 0000000000000000000000000000000000000000..ad4bbcc561959ef00375f35ce9947f7c96d0f9f1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4kawUPzXAlM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86285fe74195b8516020725bdb1adceda02d0a64d9f85daff0e852ccd4dcb871 +size 187611 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4m5RneU2kHM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4m5RneU2kHM.flac new file mode 100644 index 0000000000000000000000000000000000000000..374b85e1712b40abe9eb5b0faa05b9f60b33f847 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4m5RneU2kHM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de1203c3f175a72195e2ac1834e683787bab587d3fecc336aa4b9c5d24053314 +size 184681 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4msTzxbPaSc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4msTzxbPaSc.flac new file mode 100644 index 0000000000000000000000000000000000000000..af6e3fd11713a4554a73d4485927469668222d23 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4msTzxbPaSc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:daa387ea35c8d0a13e86540da670240a6f8962019e8e477248e8011bb78f4b4d +size 240269 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4pGvBK7Z9t8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4pGvBK7Z9t8.flac new file mode 100644 index 0000000000000000000000000000000000000000..ba3562482ae27d3d69568acce98e0301a555ba58 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4pGvBK7Z9t8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:421d43531e28d859992804aca005710b4f921a2a9f7b3fe55918e16815d511bd +size 241381 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4ukueAa9oTk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4ukueAa9oTk.flac new file mode 100644 index 0000000000000000000000000000000000000000..7daabc61ae63b52601f2635b162b1459d4c5050d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4ukueAa9oTk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4ca4fa9251a48ba7b77c2520bf044b33b94ac50de9a53742e0b79bf443c84ce +size 221235 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4v_R71cDy5Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4v_R71cDy5Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..ae389347622294be3844316c17b21b1b16174394 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4v_R71cDy5Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d0a1f7961f10ece2abadfa553a8fcffa5821dba4343c4310490ec15d53fea37 +size 213177 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4w16MdaZQpU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4w16MdaZQpU.flac new file mode 100644 index 0000000000000000000000000000000000000000..ea104453ae054da2a6d72032aebb4a20dc9dbbd6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4w16MdaZQpU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0d0b8964955a329f731430f779ae268078d31d4f7510277e3eb7764e676d8a0b +size 207613 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4wjKtAs9aMQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4wjKtAs9aMQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..b30f76ff5f0ae3a7b79028c0f9cdf2242b70abd8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4wjKtAs9aMQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6c4bbd7d5d56083917147b83e18b0d680710efcd9056eab1573216f7c20eef8 +size 266158 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4xgFvYLnovM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4xgFvYLnovM.flac new file mode 100644 index 0000000000000000000000000000000000000000..c8dcd07732eec9e3f3f761ee3a744fc044390585 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4xgFvYLnovM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf988198c273c250838c8dc1f0702e6a078ce5d4e993bcee6f6854205a45b2b7 +size 215533 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4xrL4TSgHwU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4xrL4TSgHwU.flac new file mode 100644 index 0000000000000000000000000000000000000000..6c9601dc57e1d7281d90d8f52167b8aa48a5c9b3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4xrL4TSgHwU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:76ed6c3a34524d2abbd6d93501bdcbb8b35619aa9dc23bdd6c8f7cd3b5cfdb39 +size 268582 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y4yv--RtWtHg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y4yv--RtWtHg.flac new file mode 100644 index 0000000000000000000000000000000000000000..d4a62ca45e30a8c1c234fbe455407fcbc9a82f60 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y4yv--RtWtHg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17628d4f5638963927dee8c5883a5f8bdd498f9a66715cef34c9e2a9d886df35 +size 290496 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5-4pJcg-N6s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5-4pJcg-N6s.flac new file mode 100644 index 0000000000000000000000000000000000000000..fcdd24464020b6b10b6e9e72a1aee136275d9cbc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5-4pJcg-N6s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e31b359f1c14103b28e01c1ff79edbce729609fa3481eb65ed4d8658239e66c5 +size 163937 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5-PJ49JuuRQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5-PJ49JuuRQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..eba348dc258f2b1eeb7906f4793e27dde8aafd02 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5-PJ49JuuRQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53113e706e1df537bcb259961c51f775d9e0e55f8613ec355f247879887628d0 +size 253979 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5-xKnAigQnA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5-xKnAigQnA.flac new file mode 100644 index 0000000000000000000000000000000000000000..5047ea33a5f4c7f559bbbd6885f59e04fcac23fa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5-xKnAigQnA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:035bd5ad9b6a1f661215f388dc6757d6b204d7fcab854333c7bf469d5a8b757a +size 246882 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y513JkjMF_Os.flac b/Foundation/Audio_Grounding_AudioGrounding/Y513JkjMF_Os.flac new file mode 100644 index 0000000000000000000000000000000000000000..7faba98f6e292df4ae98d5d219d843f7d7f1a6f8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y513JkjMF_Os.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ae4a324a79fa8fd1231143e36313703cfb18af6cc9e468f2edc0476cca446f6 +size 209273 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y55NFpupN4FM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y55NFpupN4FM.flac new file mode 100644 index 0000000000000000000000000000000000000000..7fe63299d2471a0434e8e51147a638eb57fc2330 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y55NFpupN4FM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:269b080c7281f99470df776cbf1284acacb530f8eef84d67cd7c1f5dee811d4e +size 222930 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y56wmOXDjjZY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y56wmOXDjjZY.flac new file mode 100644 index 0000000000000000000000000000000000000000..87b1c1285a549f42772ca1e0a1e52fc750a45ac3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y56wmOXDjjZY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f33998c6b1a51172a5de9030728ae050d6fef23997b2f71ce9669bcaf28ac553 +size 261371 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y57NefYJLTE4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y57NefYJLTE4.flac new file mode 100644 index 0000000000000000000000000000000000000000..ded8dd20a51a3144b768ccab1fe6ddd390da8384 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y57NefYJLTE4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1239b7160370bb82453da38d3da6421536116f2e5ae80542626f7b8fc09d395 +size 271414 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5ALyWjiYMEo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5ALyWjiYMEo.flac new file mode 100644 index 0000000000000000000000000000000000000000..1dc4da799ff59371ed9b5f955f4fe6e0df268827 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5ALyWjiYMEo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b72b49591a98fda21bbd70976138c7e3af682104ff9c2c4d0f86d02615f473cf +size 183867 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5CVI9p6nNuQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5CVI9p6nNuQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..b1d4448fb5b2573c1587d6f491b7583c8647f0f8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5CVI9p6nNuQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2847a5ae57c5ae372acccc2b84e2b80801f51b872ccc928c9cdc947055c89b2 +size 188000 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5CpDdfP4Yes.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5CpDdfP4Yes.flac new file mode 100644 index 0000000000000000000000000000000000000000..b69ebc9a7e10d6a547fce749ac520e947eb62dc9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5CpDdfP4Yes.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d1538945efae0306123fdf1ba50d62d257b00c038a18f150358e1fc3223ecc78 +size 194334 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5EFYHBn7660.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5EFYHBn7660.flac new file mode 100644 index 0000000000000000000000000000000000000000..26e1cd44976301d5667f454242a2f85e812de383 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5EFYHBn7660.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d72d9cc23f3cd3bee1dec0567d67c7751dbe703f1aacb8e2a2ec16621b7fb0ad +size 278538 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5FoC4i8ZbfQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5FoC4i8ZbfQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..e4dd463b5a8384413cc42be064979e3a61b9a8ca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5FoC4i8ZbfQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9bfee62118d22bc3ecedbc556f7341ca0260db4c1d0633596a8909fb4e2e87d4 +size 206597 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5GlL2iqv-I4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5GlL2iqv-I4.flac new file mode 100644 index 0000000000000000000000000000000000000000..96511849fe894782e701ffc7897a9ed62b94424c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5GlL2iqv-I4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4472d19930fff530980a0fd8c9a7a6747612f2fdf4efe6335d2c4b23f5e45f76 +size 244188 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5HRNrvHJ5KY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5HRNrvHJ5KY.flac new file mode 100644 index 0000000000000000000000000000000000000000..af2f13d3c3a392af11da4e20c4043e9500d563ab --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5HRNrvHJ5KY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7e6b4b87a2f091cdd979299547df676e82691baf739dbfb081fe5803df3d9f6 +size 228901 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5HdTmxH9A-k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5HdTmxH9A-k.flac new file mode 100644 index 0000000000000000000000000000000000000000..06e093173998b2441fa88040e6ebd3b796580366 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5HdTmxH9A-k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6286661e0126c140eaf23808f787a1237dd1589dc1e7a31dbd7e3d5aae279697 +size 215972 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5IdhdVtHUVY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5IdhdVtHUVY.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5266be89bb6872c57947142f8214152df53ffa7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5IdhdVtHUVY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e02c4d8df16fb891da2dc49b899df7f996ac5e48b25d073f51a23b47a108c1a +size 265826 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5Iz0yNdnUWE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5Iz0yNdnUWE.flac new file mode 100644 index 0000000000000000000000000000000000000000..bfbfa839f895ce534775248178535ad13407fb49 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5Iz0yNdnUWE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a957c07e69b6ca0b097ee648db5d56fc2be72e353e2f60b6dae3a08acaa4855 +size 247120 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5KmU6t2Csi0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5KmU6t2Csi0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d13e1d36f952f43de48ebc6b6c451bebbbb5f97c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5KmU6t2Csi0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c2319b38a98810add6047826f0743f5e48778a680a825a0b4e7b309c89455ad +size 261752 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5MHiwsp7Dd0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5MHiwsp7Dd0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c2310d1b501c789faf5f89e2508137f4d8b86c87 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5MHiwsp7Dd0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22383003791c7377003f14e57edd74f4a208d1304f0ac83a74e69df7f0054643 +size 232832 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5QYvAqQaDoE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5QYvAqQaDoE.flac new file mode 100644 index 0000000000000000000000000000000000000000..67467bc54ce14c26abca6ec660a0fa3927210f4a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5QYvAqQaDoE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3de28950cc6a3f508f0b300ecbc768c84e6bc7304a9efa3c87754a7cc6a2c004 +size 224803 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5RklOICz5eo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5RklOICz5eo.flac new file mode 100644 index 0000000000000000000000000000000000000000..62fa89bf54b5067902890aa67dfe914e98614403 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5RklOICz5eo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73ef7e79424bb83712fac66dfcfc6b930df7264f0bec8e5a33cc48caf57e98f8 +size 242162 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5SFxo4VFhBM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5SFxo4VFhBM.flac new file mode 100644 index 0000000000000000000000000000000000000000..814d79dae1aab519a6ed1db9f54b9690d6ec3bc1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5SFxo4VFhBM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:63699b88d8e060df0d2dc85444e08571f96f0c6f9c29e225d1b40f720783575a +size 203392 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5T6asFC8WDI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5T6asFC8WDI.flac new file mode 100644 index 0000000000000000000000000000000000000000..33163fd40cd03390aabca7a85fbdbc27fb4bc616 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5T6asFC8WDI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:996a41228ac490dff98cbf0f899f623e073e6f17bdb20df0b1a433037db4d46a +size 223223 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5UYqZWCvud8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5UYqZWCvud8.flac new file mode 100644 index 0000000000000000000000000000000000000000..ceb8794e76b31da12eda1ae9f478149f85020f01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5UYqZWCvud8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a93e7c2c8492f87eb72a33fe21819c97454dff830c935e4ea040edcb271d704e +size 207726 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5YXlfyojVcg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5YXlfyojVcg.flac new file mode 100644 index 0000000000000000000000000000000000000000..489e3bfd696fbc1ad562c7756a4476659c7219a5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5YXlfyojVcg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f9e2feebfd0233a765393d7b6358dc28ae4ab98fd0c488466f887e4454673a4 +size 251341 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5cWOyow_Drs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5cWOyow_Drs.flac new file mode 100644 index 0000000000000000000000000000000000000000..10b9982282cf4bea60a3e582b64f559d415721f7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5cWOyow_Drs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c325386e29feb4ae17dea32fd4be6a7fb464d3e7e242d3f16481b80b39cfcdff +size 182754 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5ckHyaUmfEI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5ckHyaUmfEI.flac new file mode 100644 index 0000000000000000000000000000000000000000..4dfbc5e99028b67b0abe6cb95af9cb062d3b6233 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5ckHyaUmfEI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1ad40e8045479e7fc542f70f79e6359d271ecbcc282e715b439d67ffb594d70 +size 230825 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5fihLVMJdtc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5fihLVMJdtc.flac new file mode 100644 index 0000000000000000000000000000000000000000..69190c1f580636349ce2a8544a5829ecbe1286bd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5fihLVMJdtc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:361226a2133a3e35376e6e002f88b38e74c0e576d3dca6f016f0247e8b9fc285 +size 210833 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5hxPAI1jG3s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5hxPAI1jG3s.flac new file mode 100644 index 0000000000000000000000000000000000000000..3f974acd352e2283be511ad4cab07b50ff3c1333 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5hxPAI1jG3s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ee1281d5b518b02fb535776f27941e63c1fc37985ecbf04487180ea56b3a4618 +size 171261 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5jQKX0Kmv6c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5jQKX0Kmv6c.flac new file mode 100644 index 0000000000000000000000000000000000000000..31756a71572aa2af6b3dcf8fb2ad839fa9797632 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5jQKX0Kmv6c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:285c32cece4c8c1d975013a5e7f91609b75e20d813eeab7e4b99173f438fdfc6 +size 262751 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5jpR3i8moAE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5jpR3i8moAE.flac new file mode 100644 index 0000000000000000000000000000000000000000..2554393e086fa4cbee3addd8047aae220c97e1cc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5jpR3i8moAE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32c8715ea98c30e6e47c5667ac4363bda6c78df09d8df8e560f5b5809fb4a61a +size 245476 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5kiqvp63bTU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5kiqvp63bTU.flac new file mode 100644 index 0000000000000000000000000000000000000000..460c3c25beb3fbc4299bf0b9348dfbda92ffd6a6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5kiqvp63bTU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95cf8bbb22e74d0f20ddd477593a2cf9e38946a8bffafae1fc84978fd79a4f3e +size 159211 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5mIHonRMBE4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5mIHonRMBE4.flac new file mode 100644 index 0000000000000000000000000000000000000000..31dd3e4f55fb4157455ec91a10e1f476bd23d82f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5mIHonRMBE4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:914101fc3c202654d9af611e1a0354ca162f9e9acecf6eb88ef3f0e6a134e0f3 +size 79986 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5ndGFwBfDVY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5ndGFwBfDVY.flac new file mode 100644 index 0000000000000000000000000000000000000000..29065cc3340819f91187f747f30860e0b6af1f05 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5ndGFwBfDVY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9bb715fd8689d31931bd48508f5f140553829a450ea7aabd581227634ec7add +size 237995 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5rdR6yhKU1A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5rdR6yhKU1A.flac new file mode 100644 index 0000000000000000000000000000000000000000..d99e10e60b636534c70832befc6387f8479886c0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5rdR6yhKU1A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd6982fb2f88fce129a085f9f4a96bb95351d2d38b234c670665368e567b7a6e +size 270519 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5sAeoy_rmAg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5sAeoy_rmAg.flac new file mode 100644 index 0000000000000000000000000000000000000000..7d83473fada41d2e29128bf39f25840e134428b2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5sAeoy_rmAg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:056e1d37adc5b0034916d1c66c2543173eba16bc25d941dfb8e2ed6600cbfc00 +size 286912 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5seNyvFJ_X4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5seNyvFJ_X4.flac new file mode 100644 index 0000000000000000000000000000000000000000..1d450cd64fad216bc5bd7d22db89bcd065a837b2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5seNyvFJ_X4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5da5ce886e7abcfcdf8e2d22f5e1511f6f30ce0f73998772c42fbe210cd0d9b +size 273536 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5tOUOfmpL0c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5tOUOfmpL0c.flac new file mode 100644 index 0000000000000000000000000000000000000000..4d16600e85ed633b7acdba69cc22d45a29e809fd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5tOUOfmpL0c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:856eec7d389f154057441bd3acf067294ec4db2f23eba914285f44b8bbf80af8 +size 193634 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5uWwl9DrpW0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5uWwl9DrpW0.flac new file mode 100644 index 0000000000000000000000000000000000000000..56efecaca70a2c59ebecb774b540d1e7ae04e016 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5uWwl9DrpW0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dfd0278916e4543a4cddc43d55e0457ffce0c9afa9f92b7690ddf3a55e9ce0ac +size 201064 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5wNBcUVUbsk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5wNBcUVUbsk.flac new file mode 100644 index 0000000000000000000000000000000000000000..b7c034b99aac503d9abfaaf9e8648df736f66ef4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5wNBcUVUbsk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fecf7c97d1bdcc412699908f98c4504c3a2cfebdd15b134f8fa661390394aaa0 +size 236982 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5wN_Odph5gg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5wN_Odph5gg.flac new file mode 100644 index 0000000000000000000000000000000000000000..b7c9307827d6eec4d6ded6c677e3b2f751f90a14 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5wN_Odph5gg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b07e59309bf7d26638e96bfdc4028d72f15b0b3116f949e9e4ba66e52c753a21 +size 119246 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5wfJoXRDF2g.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5wfJoXRDF2g.flac new file mode 100644 index 0000000000000000000000000000000000000000..412acd002094c269ac814c49d6cb5d9031f67e6b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5wfJoXRDF2g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:404cf6f91ba3ca967d213026f4bfae779721d917c41ff97fe25491b3d25579eb +size 205128 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5xOlEfsD0_8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5xOlEfsD0_8.flac new file mode 100644 index 0000000000000000000000000000000000000000..f3283baf9f7d9276b0871f6acb215a87d13141bb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5xOlEfsD0_8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7bb3210c3e7ba82aa2f650ba81e0670c4391e20c1e5e2c87f8264ef743888099 +size 190060 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5zPgffCgLk0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5zPgffCgLk0.flac new file mode 100644 index 0000000000000000000000000000000000000000..18a26326226fe0f9e1644bc689e8c0e1c8fc436b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5zPgffCgLk0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca1008c6db5c582811681b31b531653693976c6500c1e968fd213bc378e558d4 +size 232709 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y5zcCl0Lxt24.flac b/Foundation/Audio_Grounding_AudioGrounding/Y5zcCl0Lxt24.flac new file mode 100644 index 0000000000000000000000000000000000000000..d0e37c2d1692f58c29d3c7b006736e785d4ebde3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y5zcCl0Lxt24.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f88c3343477954feec78217824f4ea06977e9c81a5cad79a920a3f656026312f +size 230828 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6-UwJ4gAjO8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6-UwJ4gAjO8.flac new file mode 100644 index 0000000000000000000000000000000000000000..251203bbdb4ce1efddd88585d89f7894c6c0afe8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6-UwJ4gAjO8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7d70f0e61b3d3eb91854d7ab8f643debbfd5ab1e35c80c09f6567a29d94931c7 +size 250608 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y62nf3wIQrS8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y62nf3wIQrS8.flac new file mode 100644 index 0000000000000000000000000000000000000000..9154ed1e659596489f8e29fbee0b482e44c66e73 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y62nf3wIQrS8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b991efb6e4b55ed58309f39620a160bd4e7aeb02f52884112d2e98a18cb851ad +size 247381 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y64EcdCriKFU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y64EcdCriKFU.flac new file mode 100644 index 0000000000000000000000000000000000000000..452868ee99e11f0fd395f155177d4fcea26df815 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y64EcdCriKFU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8bfe09feeb8c751a534f334d98300776e31827bdc365c2fc37bb4f8a3b5df0dc +size 196136 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y64FeLMlvRK8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y64FeLMlvRK8.flac new file mode 100644 index 0000000000000000000000000000000000000000..3cb1b48b5da3f20259c56469628bf468cbf4b8fd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y64FeLMlvRK8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fadb5da2abfd430fcc836335e690a3210bd2df871b048857e3243f03b17b48f0 +size 206023 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y656Qt18XG68.flac b/Foundation/Audio_Grounding_AudioGrounding/Y656Qt18XG68.flac new file mode 100644 index 0000000000000000000000000000000000000000..ec0e69c0d375ee94715bfb94e6e47d4e976c727f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y656Qt18XG68.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ba6edc659d1f811539135af8cad8496e07e2208e6285debfc5211c4ca156458 +size 123039 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y66YGQBFkhEY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y66YGQBFkhEY.flac new file mode 100644 index 0000000000000000000000000000000000000000..41769502a01ba1b11c2cdd10fd9728d03a792e66 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y66YGQBFkhEY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd34655f7cdde5cf337ee2ac65862ac207b45806978efb321f7175f68b16a239 +size 261366 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y66cUgRJThiw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y66cUgRJThiw.flac new file mode 100644 index 0000000000000000000000000000000000000000..b26391457d187454a3fe3a554ca4e3ec5a97f7d7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y66cUgRJThiw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b78e9e43155019791c6a64dd8fa944927ee664da960dbecfe45b3c961cf39d93 +size 203204 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y66h0PEq4EwU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y66h0PEq4EwU.flac new file mode 100644 index 0000000000000000000000000000000000000000..54fa436deb1bc0a7644ec59560bb7930a90150c7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y66h0PEq4EwU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f768be308219e43f953bf3e0e873621a2cabbbc6ad010d65fef55b61ce84057f +size 214029 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y67NYH3dKEd4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y67NYH3dKEd4.flac new file mode 100644 index 0000000000000000000000000000000000000000..83e34fd02f234395c44bc20bddbd098d52e8bca0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y67NYH3dKEd4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74490bf502d0dba7488be72e7103881387476af929eb8ad1ffc7598135a0be75 +size 170711 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6F9yEqnd6cs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6F9yEqnd6cs.flac new file mode 100644 index 0000000000000000000000000000000000000000..c582248ff50db0bb5ddca3e62d8e7902e24a8b94 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6F9yEqnd6cs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d8e56b81c66f71e0b18b5c9c80b2f8043cc9c57666e82c03fa5b2ce18a28335 +size 172815 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6G11Ve9d1MQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6G11Ve9d1MQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..a66275f234fb2d522a8a62e9ffb6b470c00141c3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6G11Ve9d1MQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc5c5affa0a0a993c2b2e21c1ae7987c4deb843152159ab7f4ed1ce875f59a9f +size 215112 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6IzqjDGIDao.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6IzqjDGIDao.flac new file mode 100644 index 0000000000000000000000000000000000000000..9166c9bec4c5aaa331a8ad3469b418d5d7dcd36d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6IzqjDGIDao.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d40e93c3bfc1a6c8c940f8379498c6d5341bce68a21b730ab6c6fb1c189f019c +size 151719 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6Jwm7Uz2K9k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6Jwm7Uz2K9k.flac new file mode 100644 index 0000000000000000000000000000000000000000..4e76ea960b7b99c25119f51467fa614210d6da3c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6Jwm7Uz2K9k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0c66240a57d434ffd8927632a7b1684c08e05a6ccdb549a5f952b880d0ca0e7 +size 185447 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6K4wRfU5JEQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6K4wRfU5JEQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..4c076cc90c1d1109bf9ee7ed795b6a4e4053684b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6K4wRfU5JEQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec95d8758b6ecfb4188bd380839d07654ae1248d870892385d2ff3b9e785d30b +size 267889 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6KINFsRbXYo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6KINFsRbXYo.flac new file mode 100644 index 0000000000000000000000000000000000000000..8e2f2d40ae17c367002d03055e43a3e6c7cdbd51 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6KINFsRbXYo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:713e4c816fbe9c1575e95273ab45f49f19df6547ecb2427a13a02d853b5302bb +size 215051 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6KRIQDFUeDs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6KRIQDFUeDs.flac new file mode 100644 index 0000000000000000000000000000000000000000..00109bd5e0db3be611bccc5517c2d94107e90519 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6KRIQDFUeDs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e5eb492aee5c2b98f322797900c000ac6ba6d64dc6746afe541290b2335fa79 +size 154946 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6M6bhcnYLLk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6M6bhcnYLLk.flac new file mode 100644 index 0000000000000000000000000000000000000000..6b70ba3b99699063842907a030b67a495ef8ca35 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6M6bhcnYLLk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:13397928ac8a4f4470f1378ddf954121b5ef9f8d18d58d648d52e177f0709feb +size 201624 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6NpPjovJ9j8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6NpPjovJ9j8.flac new file mode 100644 index 0000000000000000000000000000000000000000..35499f3a1d38a0676e55a1a956bfe9b9ed8e6853 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6NpPjovJ9j8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08c4c15fb7f7caa01fab7821d787c3225cf537671366f0d99e5002930151c2d6 +size 195338 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6Q7JIacR29o.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6Q7JIacR29o.flac new file mode 100644 index 0000000000000000000000000000000000000000..47951ca32119ef8d4986f2a1e8196ba4b465dae9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6Q7JIacR29o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:783b17b640e963120318a0a860944bc533dbedbf42fb95fa3e0ab0bdf6381584 +size 243713 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6SxVYRx-g80.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6SxVYRx-g80.flac new file mode 100644 index 0000000000000000000000000000000000000000..dbe0afa94dfb46b838b808bf6354ac3157b36c7b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6SxVYRx-g80.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80195500ec1d74b1046bd0716384adbba008c6ed2ad08a308bf6a6718001abeb +size 225406 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6T2jEKesKXY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6T2jEKesKXY.flac new file mode 100644 index 0000000000000000000000000000000000000000..ef7a841044ae9bd65fe8f5805b3e2efe24fa6439 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6T2jEKesKXY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b62878aaaba313268ac90cfb83eb2ed78979adfdedb2021d038b62431ab78155 +size 259055 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6TP8w9CcdHk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6TP8w9CcdHk.flac new file mode 100644 index 0000000000000000000000000000000000000000..c81afee11658d5cf38146342939243b84605cf91 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6TP8w9CcdHk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:458d2f292425256d8e2a203dd96364db51e81aec85e68eaee29d72957dc75579 +size 284355 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6VJWBMh416c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6VJWBMh416c.flac new file mode 100644 index 0000000000000000000000000000000000000000..1c0d4d469a9846bcb38b9103bf15ae930cae1bf5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6VJWBMh416c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3bc2e12fc16772e5ef17a988baa1e3bb40c9afda958fe654bf39d7f1a95a3b76 +size 167496 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6WteI_2ipko.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6WteI_2ipko.flac new file mode 100644 index 0000000000000000000000000000000000000000..9b1d88e5ad9c39b96eb3518435066cfa9284aea8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6WteI_2ipko.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b476270e7d191364ef7299d2379fcc416276759977c9f1c6ad650d64360d2495 +size 236618 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6YKDBuntOF0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6YKDBuntOF0.flac new file mode 100644 index 0000000000000000000000000000000000000000..3322522b26ac3e6fb98044a8236cdadb07d0a18f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6YKDBuntOF0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ea966ac5cbca29563ebfcf059df29b07332d19e8f75c7199c9b1cb4c56a984dc +size 150501 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6YWK8S1thvk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6YWK8S1thvk.flac new file mode 100644 index 0000000000000000000000000000000000000000..15ac9694dc484c9ff4d925bb3323cf414315f36c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6YWK8S1thvk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:039d73408e4fb6f4778715b4a007530f1c838b252f67191ca2abcffa5de8f262 +size 199743 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6YX6esYQEDY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6YX6esYQEDY.flac new file mode 100644 index 0000000000000000000000000000000000000000..701c8a9ddaf961ece8e1e670834ec4dfd9632841 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6YX6esYQEDY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eac751cfef58b01d6ba371e6d07b610ceac6f64980d7e56bdfa52240fd32bc3a +size 191032 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6a97tMZJNEA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6a97tMZJNEA.flac new file mode 100644 index 0000000000000000000000000000000000000000..dc06ce8bdb2364363eae3f0cb6d4e002b30ddd6c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6a97tMZJNEA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94c2ccf6f9c4c8c3b18300feb62f6702fb08063f7430ad83826465492c7803f5 +size 207170 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6aiLRDp5hco.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6aiLRDp5hco.flac new file mode 100644 index 0000000000000000000000000000000000000000..154c0c3fb9ea80bc32758832f07fd9031ba9710b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6aiLRDp5hco.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a7fe737f0ae22e07a14ad46a290a7447eaa35dc22b54512ebea0c9919a1e630 +size 138465 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6cS0FsUM-cQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6cS0FsUM-cQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..c9a57d0d8acfc98ca9148584eecd19a9575d812b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6cS0FsUM-cQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae39112f6af7af85666dd7ec60374277147c784d4f7d22fc6517389ec5a55789 +size 215546 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6cWIWgtMLjg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6cWIWgtMLjg.flac new file mode 100644 index 0000000000000000000000000000000000000000..83745fa37c722eca1fec88c1647c42326e2960c8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6cWIWgtMLjg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0faab57496a5a6fa82050cb76ce0989c0536440e6d4798c95d1ef83432819af9 +size 200980 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6dLV_94FiNM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6dLV_94FiNM.flac new file mode 100644 index 0000000000000000000000000000000000000000..979ee0fffcb8389a07c65bb656f3eb0165e0dc08 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6dLV_94FiNM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:79cb5f62e5fff39ca47a514c9a526d8a1fec78beae19452705bf012d257867dc +size 243066 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6eHnIvL0N7M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6eHnIvL0N7M.flac new file mode 100644 index 0000000000000000000000000000000000000000..29fd85867090935561760966b89bdf229081f202 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6eHnIvL0N7M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:deedde74838eca00127784161bc4434add47f744e7375bc12e0a4a201bcdb15a +size 243884 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6gisYWJffTI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6gisYWJffTI.flac new file mode 100644 index 0000000000000000000000000000000000000000..843f41f2542f52b438ae2e114a5d048283eee657 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6gisYWJffTI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c08ddc5f52c0f4fc0b7f77e0eed90b6f3ef7c1680396046080b7c0390145780c +size 288617 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6h4hZiywOxE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6h4hZiywOxE.flac new file mode 100644 index 0000000000000000000000000000000000000000..f5bc3684ec093c6f04b0d9ab6d1f04bb49acae13 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6h4hZiywOxE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2830ac05892a197f0556db83b1389052d91597c6347b192d8c21fbb9d0b92b09 +size 201285 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6hRncnuF1NY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6hRncnuF1NY.flac new file mode 100644 index 0000000000000000000000000000000000000000..26fbc9161b31fdd16fc839d755c1649a0a0f2ed5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6hRncnuF1NY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7f14a56ca09a8b57ec3d104e2003afb04c8443b41a49782f5bb29e5a53f6e29 +size 258386 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6i-BHYm4hYI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6i-BHYm4hYI.flac new file mode 100644 index 0000000000000000000000000000000000000000..4b807690fc9770189a50d257be8d0dd1ab5c4673 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6i-BHYm4hYI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64a75727a7dde81fcbd4ca85179a55da6cd155ce8b93e3ae4b8ee21b684f5d1b +size 149065 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6i5StsTUj7k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6i5StsTUj7k.flac new file mode 100644 index 0000000000000000000000000000000000000000..178264755b910f4409b468c42006fee644c973b9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6i5StsTUj7k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f403333813a727d3ea17b0e3e5258bbbd7889d2f835343e19bd5b3f2af23c297 +size 237042 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6ipK0W9CbK0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6ipK0W9CbK0.flac new file mode 100644 index 0000000000000000000000000000000000000000..fe8a1d3b60faeec26d3252885d70830beadad1e8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6ipK0W9CbK0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78e1fc171b112d2a388b674df9917e05c2399a7a44bd3c9d35942249598ad818 +size 182204 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6iwOOiYQAt0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6iwOOiYQAt0.flac new file mode 100644 index 0000000000000000000000000000000000000000..f559846d8b20f49339f66dd5b14dc83f750c18d0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6iwOOiYQAt0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35d6cafe09b0be9d42b2412afc4e08d3d1e98d306f2fbfa9fcd02e858949612d +size 201056 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6jFl0KW-S84.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6jFl0KW-S84.flac new file mode 100644 index 0000000000000000000000000000000000000000..5aa1024d6e758efba11da31c57d55c52bb1bed84 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6jFl0KW-S84.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22c994194555a6dbb74b3a937344a15e2c8bf610bec3874e19eb1922955a1bd2 +size 231030 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6jLu70WHfEU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6jLu70WHfEU.flac new file mode 100644 index 0000000000000000000000000000000000000000..824db6a5d0973a9664d965477d4e2d80612acb8c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6jLu70WHfEU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74c850ddd7c0e99470cae8c28b53f37f491f153025c11d8a7027c85ee631e008 +size 156589 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6jRpHlZgVy8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6jRpHlZgVy8.flac new file mode 100644 index 0000000000000000000000000000000000000000..917fcd0449d20ab7c7f0c4c32cfc1a3b497e1e77 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6jRpHlZgVy8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e506b33be35f486650e9384462173878288342ab12369a6472d8d5b66b99a96b +size 137061 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6jrffMFpYpE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6jrffMFpYpE.flac new file mode 100644 index 0000000000000000000000000000000000000000..3372f7a2a7637df6c5b4a7677351f4f84f3a55a9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6jrffMFpYpE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a91e8e81bcf5f59b2ee0249acab1054c4abc8bd412d6bf8f235832ff7caeb3f +size 183022 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6nPirCzxDV8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6nPirCzxDV8.flac new file mode 100644 index 0000000000000000000000000000000000000000..7f7488299dfb66be22ee95c227caa3b58add063a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6nPirCzxDV8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c8d5ba08fc643f988c12903747ff01eebd3b21a1f323892c5c5b44f1c88e02 +size 146702 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6nr3SWNml9A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6nr3SWNml9A.flac new file mode 100644 index 0000000000000000000000000000000000000000..638c8daa3526fad36785b59887cdda0e2693df4e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6nr3SWNml9A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ec3dd9547a4f21de265f442abd4708138f0eba5c4d4ad78d7d69ce5d9ea8a84 +size 219883 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6ohU5qMW9yo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6ohU5qMW9yo.flac new file mode 100644 index 0000000000000000000000000000000000000000..bf5fe0c7121a9c59f9632182ae894bbdd34f7aed --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6ohU5qMW9yo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:678247053764856ee6e9b0a924b35284e2103451042d4108b07424b990fae1c7 +size 193567 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6onVPeqqNsk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6onVPeqqNsk.flac new file mode 100644 index 0000000000000000000000000000000000000000..1798e6c58d7eec1c3f63fb97f6df8bdf090c8731 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6onVPeqqNsk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:200fca731953a47394ab2187fe3825ac20ab24973c5898f40310d0e616439d4b +size 221490 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6qD9LHCKbJg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6qD9LHCKbJg.flac new file mode 100644 index 0000000000000000000000000000000000000000..70f6f19badfe29fecf220a5f6bdf4802f4593bc4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6qD9LHCKbJg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48d7db4ff4d145ca6c81b4a991195efcf038c6d768e89995e4be65a90e86d510 +size 208101 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6qxcYCOkw8I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6qxcYCOkw8I.flac new file mode 100644 index 0000000000000000000000000000000000000000..61cf23401a91c468e8620dd02a56660daacd4506 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6qxcYCOkw8I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f0676c4e22c93090455b0f88889cd9e7eeeeff19663679dba2d343b8004ff78e +size 151768 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6sFsyoK8Nn0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6sFsyoK8Nn0.flac new file mode 100644 index 0000000000000000000000000000000000000000..b36f3cc5aa4d1abd02de663d2957df90f3939a7b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6sFsyoK8Nn0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fce35c22088504fd1c2bc2733e9ba98f67afe7e21c3d875a507424a6450e393 +size 233505 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6sXAXmSQ5To.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6sXAXmSQ5To.flac new file mode 100644 index 0000000000000000000000000000000000000000..a2c0d52a6fb47140393db8ab9e5e151bd57270a5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6sXAXmSQ5To.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5544fb315730df9b5eb3309800afd13da06aed77dd375d41613bef0890756c3a +size 186503 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6t9R8V0HNXM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6t9R8V0HNXM.flac new file mode 100644 index 0000000000000000000000000000000000000000..5ecf2afa8986c2b191ea9521727fa20d740c71b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6t9R8V0HNXM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f8243c9728b0dce1b644e14e514e7a35c3fcb8fc25bd38269085d1a71888ba9 +size 127078 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6wtUZBZIjBs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6wtUZBZIjBs.flac new file mode 100644 index 0000000000000000000000000000000000000000..bc7f337f7d6e5aa8fca9ed8ab77e51245c938f68 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6wtUZBZIjBs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d65e0bc19d0fcd030910ee7413defbc7e448b49f95379d9d14e6e667fa100934 +size 204686 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y6xpurMt7wo4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y6xpurMt7wo4.flac new file mode 100644 index 0000000000000000000000000000000000000000..9ecbecdd321682630de8038121e7588a0c4ae89a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y6xpurMt7wo4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65ab787ac0978aa4e687a930d6187f80c7dc63c92802ebb53bb370c391f2beed +size 159681 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y71iXgJIvlTg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y71iXgJIvlTg.flac new file mode 100644 index 0000000000000000000000000000000000000000..012ce38087c7fd95e06f3544253816d6c9dfe760 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y71iXgJIvlTg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f661b47997f5c2c2aef7f1bb0a992a5db5f0bdd9eeaaed81d8cdc837807fe642 +size 185347 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y720-k49ehJA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y720-k49ehJA.flac new file mode 100644 index 0000000000000000000000000000000000000000..244e75e963b1f8b8291a87b668b17f44ceca8b05 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y720-k49ehJA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:909cd0783cbf9391678917902c9d772115228667e4b1ab7d17171f65c87d25a5 +size 188994 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y72J8fjETl6U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y72J8fjETl6U.flac new file mode 100644 index 0000000000000000000000000000000000000000..7699963f824018f9f18b367af2bbd83e4f28389a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y72J8fjETl6U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67d95fcf65cd5bd22e6dfb4876e270edc7d7fbd65bd41786d263b214e6ebffa5 +size 216411 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y72JnuHYZuVE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y72JnuHYZuVE.flac new file mode 100644 index 0000000000000000000000000000000000000000..f61a0a77867823985ace4465b8f8a04d6ef1a821 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y72JnuHYZuVE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdb369bc2b2b650c64f9ca23f7ec23f9cd83bd7279be074bcebbe202a81ffc71 +size 202688 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y736sQyNeUpc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y736sQyNeUpc.flac new file mode 100644 index 0000000000000000000000000000000000000000..8d159004c8b6c5c7b76363ade291d11b956ff0d3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y736sQyNeUpc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8a7455319e026677e5580abce8e03c8cbbfa40011a75b33b9ddccefe26022f3 +size 237990 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y73Vz2uj8iKY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y73Vz2uj8iKY.flac new file mode 100644 index 0000000000000000000000000000000000000000..c940a543c25f28ccc7cbe66eb3fa50f0f7651081 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y73Vz2uj8iKY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac19c66c5d57b2b49bd5333abf2a115061b96d91ce57bc35e5fec605f1b8d35f +size 244503 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y73jR8wkpTt8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y73jR8wkpTt8.flac new file mode 100644 index 0000000000000000000000000000000000000000..6a74852e9483d6485093728e204087d6b8e2455d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y73jR8wkpTt8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbefd7d4a36395259e9514f2f58a56070ef8e4510ad91546711d2e4a74b4180e +size 226358 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y73phKdahss8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y73phKdahss8.flac new file mode 100644 index 0000000000000000000000000000000000000000..8e5b155cc84841277af62a11590732e73eae1763 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y73phKdahss8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3dff6ff84561c9c4ec9fd6541aee942c8d4fcdd1572dbda7c43a8734c962664d +size 252035 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y78nshhXgM6E.flac b/Foundation/Audio_Grounding_AudioGrounding/Y78nshhXgM6E.flac new file mode 100644 index 0000000000000000000000000000000000000000..c58d0e8c01928d18e548a6ab15cda63262345e11 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y78nshhXgM6E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf37b8d2ca8c04160438455b142f4904c656c4bb7e81b2565d34e57695b15201 +size 208298 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7AScPytrpac.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7AScPytrpac.flac new file mode 100644 index 0000000000000000000000000000000000000000..83996ca177c38a52c33b23e5cdad3c12feec66eb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7AScPytrpac.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c8388b50e50b26d5330214c0e4065fa04238b5915197a3b22e364d173e5cb93 +size 197100 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7Ay-gVQ4oSc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7Ay-gVQ4oSc.flac new file mode 100644 index 0000000000000000000000000000000000000000..f0cf18a1284d5a8e494f3e15a797afe1d7a956c8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7Ay-gVQ4oSc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1d187fb8fb849b0903651d66ee69d214d04e0517f189cdd181ff58d23c8001a +size 210880 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7BUwVfKkw_w.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7BUwVfKkw_w.flac new file mode 100644 index 0000000000000000000000000000000000000000..b996fcd7246d81b0d01c13d6261c93fa1e4c4533 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7BUwVfKkw_w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c25e620906dbd344f846a770e17d6c18f92b6cf1bcecfe16e9b701d652883d9e +size 233044 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7Cm9bUZNVuk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7Cm9bUZNVuk.flac new file mode 100644 index 0000000000000000000000000000000000000000..2c5dd8270cbd394c4e9cb4a08cc0f87a602f2790 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7Cm9bUZNVuk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:61f511d146b276b8bf2698c41b8779e966bc7be052c77432146a81ef2e2051cb +size 305032 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7EWWhDSFreY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7EWWhDSFreY.flac new file mode 100644 index 0000000000000000000000000000000000000000..abc581e3063f1f1160e0ca82bf168dad7bfdca41 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7EWWhDSFreY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6e3ba4899a4b69fc1b0b828949edbb55d58d2963951c13d480fdf938cd01fdf +size 218254 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7GoQ8HaVAnA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7GoQ8HaVAnA.flac new file mode 100644 index 0000000000000000000000000000000000000000..801bb3c32271fe73e547a63e8a70e0c0da8c91a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7GoQ8HaVAnA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6fc241ca0fe23544e9d0e20e9055b5f49bac116899c1bbe7f3b7f4aa72eeb9c +size 300890 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7HyTIyWWRu8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7HyTIyWWRu8.flac new file mode 100644 index 0000000000000000000000000000000000000000..8dd64da7d229aa1b2f8bcf3b45be799efc209f76 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7HyTIyWWRu8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d98a6d0e9a7b7c3be8d14cd1d01203aa243256fba5b488702c1cb3d62d726dd +size 250730 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7Ivl23eqQto.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7Ivl23eqQto.flac new file mode 100644 index 0000000000000000000000000000000000000000..b8f834e32b62c1b2ca29f4b9a8c5b894e12e815f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7Ivl23eqQto.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42d84058ec915eed92ff7607029c22ad5eb5ab0e83e90389bd229a352e977f17 +size 168602 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7IzG88ix5Wc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7IzG88ix5Wc.flac new file mode 100644 index 0000000000000000000000000000000000000000..379416164d25572afca33ae238f5e91494ec03d4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7IzG88ix5Wc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5558e249ef01637b37797bb3f017be67eb81b51f2dcd8dfacc7a54609cb47bc +size 225632 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7OAWoKiUEXc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7OAWoKiUEXc.flac new file mode 100644 index 0000000000000000000000000000000000000000..70df8908548502df81aec3cfb9e6fb43c34201e8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7OAWoKiUEXc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:135ba541305e2709c94c9cb33d8c22a9c5a1e9175da3d02fdd857502dcc76284 +size 248225 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7S3Yf3QBvzk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7S3Yf3QBvzk.flac new file mode 100644 index 0000000000000000000000000000000000000000..023c79b4578e620ff46734855210e232d93d0681 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7S3Yf3QBvzk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78b5f860713d675fdd744a01e5920626ecde26dfcc32982370efc87390409083 +size 272549 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7UTiEhYDBQs.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7UTiEhYDBQs.flac new file mode 100644 index 0000000000000000000000000000000000000000..2cda9049b1339d7309a430e74f924885149a6836 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7UTiEhYDBQs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24679a6c10ad052223058d28f2d4227e8daabff2fbd3725fa1632a00bf6b1afc +size 184955 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7WP4vfMgZsc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7WP4vfMgZsc.flac new file mode 100644 index 0000000000000000000000000000000000000000..01f098cce66c3ee63fc4120537a16bc859783245 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7WP4vfMgZsc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:06697e5c4bf42f67028830110b7f2a9dd4b7eefb676e169b553a31ccd099ee1d +size 212229 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7X3i7llSK9U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7X3i7llSK9U.flac new file mode 100644 index 0000000000000000000000000000000000000000..7fa28572f1ce1902d0ccaef2af85029583646f21 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7X3i7llSK9U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e64bdb1b43bfb1d643d6eab6ca4566bdfb6e6aa016f9eb2c3baee3318cd72d44 +size 281438 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7ZD1_tPZ1qE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7ZD1_tPZ1qE.flac new file mode 100644 index 0000000000000000000000000000000000000000..a73f5342f3537661d010e013932e1295c197c6cf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7ZD1_tPZ1qE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:56716e00b9f52a733c87c923a877e9ec255830058339f851f932577859aaa2ae +size 277917 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7ZeCyUZT8Dg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7ZeCyUZT8Dg.flac new file mode 100644 index 0000000000000000000000000000000000000000..e47b7b2b5961b8d9abde3676dfd3f3552f2280fd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7ZeCyUZT8Dg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01492d5253be7beebb605d90e3d967e724082d439439f728838d4e9269b46f9f +size 294916 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7aI9LqS4jK4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7aI9LqS4jK4.flac new file mode 100644 index 0000000000000000000000000000000000000000..f68b8dd2a7ce37e6419fd98d2a606b303dc6a3fc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7aI9LqS4jK4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d66befe5e5ee3e8a62da0eb35ea377cd700dbf304bf625ec26240addcc1810fd +size 193560 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7aifdCWotdI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7aifdCWotdI.flac new file mode 100644 index 0000000000000000000000000000000000000000..c625d043504444462d23412e34e72b5f66c71b4c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7aifdCWotdI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4d6c3d1f5f3eb98d67b85812a73c595d2f5af383bee00378015237ae3adcf5d5 +size 243252 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7b3s4rfWF50.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7b3s4rfWF50.flac new file mode 100644 index 0000000000000000000000000000000000000000..153a5c5ac043a3a786533b62d4d6f4d4e793dd6e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7b3s4rfWF50.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:983fa563ba946bf309f593df3ad720af75085e9aa6460380460e3b664d7f2f37 +size 231237 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7cUkRQrNSIE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7cUkRQrNSIE.flac new file mode 100644 index 0000000000000000000000000000000000000000..282008adf8aa687b664f8f6ab1cc9a869b9214a0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7cUkRQrNSIE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ccbcfca68c01bb5314acca904f138820c192123a0a48d99f026acfea351be98c +size 202509 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7dNJWByoz10.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7dNJWByoz10.flac new file mode 100644 index 0000000000000000000000000000000000000000..d12d62a8135d4c60086c1889cc0b37dc9831ec3b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7dNJWByoz10.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:74b22bd7727a20ac2a407f17dcb4914f6f4cdbdfb94ab540b70439a4f6c076db +size 139926 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7h-wCrAdSNY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7h-wCrAdSNY.flac new file mode 100644 index 0000000000000000000000000000000000000000..95b328f6706b7d1bc1e63c080cc2d33e7531ca16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7h-wCrAdSNY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:039c95d1b5cb42e791c7f766a2951ac275efd0ccafad3e8cb13b3b35be3d7533 +size 213637 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7i8tr14DaOI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7i8tr14DaOI.flac new file mode 100644 index 0000000000000000000000000000000000000000..55b4ff9285301a59e3834a5d47ecc7ee2f2037c2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7i8tr14DaOI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a04d156093a8708ed0d730cebbe7975d69f3f65dd5ef66f6e985f731e4deb8d +size 181385 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7kYXCkEGxwI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7kYXCkEGxwI.flac new file mode 100644 index 0000000000000000000000000000000000000000..136a5686b89d3f462d45faa376d0d972d02825b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7kYXCkEGxwI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c8dd1bf386873fa02421c788f01d947c9eb713dcdee8643d2ebb3cd640f5d711 +size 181313 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7li4oPDlkvo.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7li4oPDlkvo.flac new file mode 100644 index 0000000000000000000000000000000000000000..e120d1e7eadcfa5924b862bfa25c6881d7bd610a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7li4oPDlkvo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4917da07600636e4535d4a83db52ab864f90501b466db84201718f10e31b133d +size 249380 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7mQepeQImgY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7mQepeQImgY.flac new file mode 100644 index 0000000000000000000000000000000000000000..93c044eaa5b2ec10ec98247a165c9fad3fb56c3e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7mQepeQImgY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4db44ccd6bc73ad966190d9964a116bf5315f56a1452b1b0126bd5541c1fa0bc +size 239073 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7o5Q7fKYXL4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7o5Q7fKYXL4.flac new file mode 100644 index 0000000000000000000000000000000000000000..ff4638fa61d17ecd438fb1f02f72b39dddd30ccc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7o5Q7fKYXL4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e951ed66bf9651f13dbe6a23ec17217cd20e85fae137125f8dbee7b578a5e693 +size 132006 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7oqLwoYw-hQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7oqLwoYw-hQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..00d54d56ba6d6d89f6ded25caef9f4c2710d0c34 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7oqLwoYw-hQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d41a537d2e8609c1b844dfe009bbb52ab7f4ea86c1f838c32c8f1746529569ae +size 254309 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7sHIi7WZkDA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7sHIi7WZkDA.flac new file mode 100644 index 0000000000000000000000000000000000000000..860df58e7105c2e8927f9ee6075370b16268c0b6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7sHIi7WZkDA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57632731386f19bafaed9536afa3853e5e9b3206c07af53d152701a45a2808f5 +size 134902 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7wyV4mU-E4k.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7wyV4mU-E4k.flac new file mode 100644 index 0000000000000000000000000000000000000000..fbfcca5425d4869ac977142e20136402558bc273 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7wyV4mU-E4k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a76b52228ab392f5c1999e9771d541cac465796c22a0d6146f813bec82210334 +size 192712 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7xjWD6j5-RY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7xjWD6j5-RY.flac new file mode 100644 index 0000000000000000000000000000000000000000..573699f2eab30e9bad1bef2cc4c41445a5e60b74 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7xjWD6j5-RY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0e9117741436614daed50805df6a700ee56df3ccc7472f9be17631d31bf893da +size 255354 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y7ynQ9wd6neQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Y7ynQ9wd6neQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..585f4eb46761a6f2a405b5612453def51fed538a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y7ynQ9wd6neQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f41f3c6e8d479cb8088e90aa0146418633581766c58fe7ad2312b47ff3c27899 +size 266136 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y81C5ysSrqyY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y81C5ysSrqyY.flac new file mode 100644 index 0000000000000000000000000000000000000000..67f40da39355aea8ec338e6cfeb0ce68907d0d5f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y81C5ysSrqyY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abcc265186e620aaa48a7b0d462f1957ebe4704a1e32bedbefb90431d5a6bb3d +size 215975 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y83u2vOHb52Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y83u2vOHb52Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..7d88127f5ef7ef7a4aea93de0889890bf1c2d7f2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y83u2vOHb52Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d64a0baeb20634af3c1db217955c5196dd82fcaf45b48e4ba1db99d63b59335b +size 146841 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y84NFbeQy5CU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y84NFbeQy5CU.flac new file mode 100644 index 0000000000000000000000000000000000000000..bc2529e8704a5f94a4ef1480599b9d5d8601324d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y84NFbeQy5CU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0e58d2fb4f3c5f61dcd714052e660f973cd5e8575bfdf6f76e5da67cb5e4c67 +size 224908 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y880IpGRqC5c.flac b/Foundation/Audio_Grounding_AudioGrounding/Y880IpGRqC5c.flac new file mode 100644 index 0000000000000000000000000000000000000000..295b89ebcc7696454a0a198a5522c65fe09c69b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y880IpGRqC5c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ce55cd304053a61935186cb3770948cb17c9cc750479e1dd94e300673fa3a7dd +size 202305 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y88id3bEvC4s.flac b/Foundation/Audio_Grounding_AudioGrounding/Y88id3bEvC4s.flac new file mode 100644 index 0000000000000000000000000000000000000000..106136f574e39df405aaf617bcbe636d8d0b5979 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y88id3bEvC4s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6197a4c51613647b8f64140a17ce51e6152861aa64b371e4030182ae0d8e0fc +size 136043 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8DrIJDw3b_M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8DrIJDw3b_M.flac new file mode 100644 index 0000000000000000000000000000000000000000..6fdf1ceea3081b0f321e41bd856c11383ebc9946 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8DrIJDw3b_M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:efae88322a91c76fc71622bf6d5b55b8814362fc0f6b0d5b2fb95b1adf6ac334 +size 222665 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8InQJ0Nng0M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8InQJ0Nng0M.flac new file mode 100644 index 0000000000000000000000000000000000000000..0fe43201e34d4bdbed929f111a67e079d58937c6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8InQJ0Nng0M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbc2d3e823f5e284fbb71b41cc3637461125f16a61c30d6f8f595a21b08973f5 +size 180578 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8Rb3I8SyT1U.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8Rb3I8SyT1U.flac new file mode 100644 index 0000000000000000000000000000000000000000..5022071a82c80bb19b1ad987eaa41a406a365682 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8Rb3I8SyT1U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:197c150cbffd5f1b0b09388ac0b5aa1dec4d65b82280ea720ed854c50197b098 +size 232259 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8RjytDXc23M.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8RjytDXc23M.flac new file mode 100644 index 0000000000000000000000000000000000000000..2852c277a4c12964a405e2efefc68092357c0b79 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8RjytDXc23M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcc7d624c8ba80a012d05a186315c460d4dbba39bb105bf05ff3100f81c30b54 +size 148532 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8SYX5jXLm3A.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8SYX5jXLm3A.flac new file mode 100644 index 0000000000000000000000000000000000000000..62120cb28d9afc3935616d3bf4e53184e9218606 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8SYX5jXLm3A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db650b2f62b93e3e89c65895d9f78744f4f748bf379722475cfba34c2d6dc032 +size 215293 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8VoiWkzvNQ4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8VoiWkzvNQ4.flac new file mode 100644 index 0000000000000000000000000000000000000000..6f9aa430e5cb0fc1eed0b219e7402f307aaba678 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8VoiWkzvNQ4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69cf5fe07fc6cb4e233fa31095c7a529e6b6c211a977890097a042d86c0d41d8 +size 252121 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8WbUPEYjo1o.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8WbUPEYjo1o.flac new file mode 100644 index 0000000000000000000000000000000000000000..e4d33f8065e2665e8b2ba4ba4bcd3ace260e0434 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8WbUPEYjo1o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1bb0a7e070563ca7538bd6451198f5d10a71f9653cf1c340b1606749071ec707 +size 242851 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8YDpaKb3QZk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8YDpaKb3QZk.flac new file mode 100644 index 0000000000000000000000000000000000000000..bc9b9f2121e6de325824bd2d904fb7cf7a61719b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8YDpaKb3QZk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa3933771ebb0ed684f5bf903d7a923da9eddb6481b23aa25a59646ea0f6fb63 +size 161863 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8ZmSpxseOtI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8ZmSpxseOtI.flac new file mode 100644 index 0000000000000000000000000000000000000000..1f2a382ca70f0a03afeb474de805aa96615cdadf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8ZmSpxseOtI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44685449d1caabfe8e1b96b381240ec1bcfbaf21c25c83973f1e5a197fdaf34f +size 136891 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8ZqX99afNqE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8ZqX99afNqE.flac new file mode 100644 index 0000000000000000000000000000000000000000..489c42b3c44c84bcfee10e362115509e02beae01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8ZqX99afNqE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:03a8520e249d64fe1e812d7cc7de060efc1446cfe69fcfd11e8b43d5f9b14491 +size 234982 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8aDTJApiNrk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8aDTJApiNrk.flac new file mode 100644 index 0000000000000000000000000000000000000000..ab41178f5ea64b3d6d4345ea28da18b87632a7b6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8aDTJApiNrk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d46f27d81088661a5575ed38f5532027b88baafaa33c1dcfe6dacfab9759c22b +size 238895 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8fVq7MMxhVc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8fVq7MMxhVc.flac new file mode 100644 index 0000000000000000000000000000000000000000..1b347317f091411484da4aa67d99a72f1389c24d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8fVq7MMxhVc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d3cba642cca2834ba9dfa93042070f0ac3542c3590ce5eaf3ec6a89f86758a65 +size 136511 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8g8UfyxJ3qE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8g8UfyxJ3qE.flac new file mode 100644 index 0000000000000000000000000000000000000000..da28a186804ed76a5e14bcebc3f7eb3cb9ba270d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8g8UfyxJ3qE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da2d568f0696d4298f04e846917f9aa383b694edb41c3a32fe6a1167facb89c8 +size 248832 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8kDguJE2oC8.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8kDguJE2oC8.flac new file mode 100644 index 0000000000000000000000000000000000000000..b095ab7345595bd847e9c6c1969b840dfc769838 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8kDguJE2oC8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07eea65b8a2c4d3a384ee614b3ebf9da3fd42e39c858e4bd1c0d413d0c33bfa3 +size 194846 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8lhdsS_xrvg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8lhdsS_xrvg.flac new file mode 100644 index 0000000000000000000000000000000000000000..aa6af3dd21e8b7b2cf9ac273cb74b23859e4266d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8lhdsS_xrvg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a271c5a8da906e645cf93a0a633fdfd759376b69358ca216651144e36057ff9b +size 272604 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8naj2THt5C0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8naj2THt5C0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d02bdc50e21227824063011caa462e232131f460 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8naj2THt5C0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:89d1687d4f44743b626a62b47fed07b6b487f3bdb6eb42501c370d5ea4299d2b +size 195027 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8qAuo8Dmr2Y.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8qAuo8Dmr2Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..bae9df41e89a9580788f716e4f98f565b6e454ca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8qAuo8Dmr2Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ddd41f2e11397d7ccb123d18f5639acdf1b2c02b8ecdf406ff2bec2bdd4a4c8e +size 199412 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8t4PWuA8sSM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8t4PWuA8sSM.flac new file mode 100644 index 0000000000000000000000000000000000000000..1370f9dfd42a9636695c9882f2facedd3d456624 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8t4PWuA8sSM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf7db744b455bde99f6b319afd99144591cf7aa5b50db3212dd78dcbbe50ce09 +size 253122 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8tC9BD6s-So.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8tC9BD6s-So.flac new file mode 100644 index 0000000000000000000000000000000000000000..808a7a7ad90bce13b95f39138552852bd64168d8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8tC9BD6s-So.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cea1c968bf1260e5efe6d144475fd362d9d4b98899614f9c690ce5311d9a46a +size 226462 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8tskw5JW2d0.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8tskw5JW2d0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d6afb7f3e0859c14d3b8a65699152c5305f8a2f2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8tskw5JW2d0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11e14a9de81205118ea11382ce5b86d178ac943fb2bb1676f3e6b7cad60638cc +size 174416 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8ycflE3dIHw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8ycflE3dIHw.flac new file mode 100644 index 0000000000000000000000000000000000000000..0d717abbcbbb82c5d0f0794c62dda9b95dfdd58a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8ycflE3dIHw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fd4a2551f46fb624e391358cd16eaa5d1da53b3807ac5e1bbbbd146fd4786735 +size 232891 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y8z8oy5mkdt4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y8z8oy5mkdt4.flac new file mode 100644 index 0000000000000000000000000000000000000000..50910602fbb76839d8cd9ea60ee1ac2e5ff1f95d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y8z8oy5mkdt4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07af2d990d1e452bfb74760b714cc1765e5a831555f8d7cb70b73662f3f4e6c2 +size 202655 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9Dz4Sisz-Z4.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9Dz4Sisz-Z4.flac new file mode 100644 index 0000000000000000000000000000000000000000..ce11b8c357754888323c2a0a74d3ea8eec088a91 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9Dz4Sisz-Z4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac932c07e5aaf31a207b99b0db1b65bb483e9aed80e0570cb0c61124510cd21d +size 260739 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9FnSoyIhUGI.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9FnSoyIhUGI.flac new file mode 100644 index 0000000000000000000000000000000000000000..db007dcee3f26cbd362f604868b8006e68564949 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9FnSoyIhUGI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09a04baa4b6275225b69cd6d19b81fa70235e41c320fdde26a9eee63d7f8a34a +size 253744 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9IshV__iHbE.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9IshV__iHbE.flac new file mode 100644 index 0000000000000000000000000000000000000000..4dcda78a811664a9c57c52091818268240d55f48 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9IshV__iHbE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:904b63ec9d4f6a9e97b8281e3a454e16cf9d2fc0f45905d71edcd7e379c74e23 +size 183265 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9KuT6SOwobY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9KuT6SOwobY.flac new file mode 100644 index 0000000000000000000000000000000000000000..a4a3f702ebd34778d5b7beb50466f4cdb89c9100 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9KuT6SOwobY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11fad02d51bd77a5d0b82cf41634744cd938f83114e7445647e0ab002d34edd3 +size 107481 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9LQEZJPNVpw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9LQEZJPNVpw.flac new file mode 100644 index 0000000000000000000000000000000000000000..9f8d17e3b4feae9de99f67b457e3ab667786d21a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9LQEZJPNVpw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f312c26e9fbebe9907e4965b2e108484553eef7eda0a5b4b91062187802785c +size 258337 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9XtBx5HwlOc.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9XtBx5HwlOc.flac new file mode 100644 index 0000000000000000000000000000000000000000..e3abf0f34f060303942165c52e8bb73df225dfa3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9XtBx5HwlOc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23d6d4af9d4443c9b37883c9fd310d4ab78accd7d0a917762fd0d490c016f2f5 +size 238279 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9ZDBsJjzNOY.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9ZDBsJjzNOY.flac new file mode 100644 index 0000000000000000000000000000000000000000..8ceb1295c30e2ed2dce885b45d9fc7450fdb6370 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9ZDBsJjzNOY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5593b69fc1900fa57af6d2b2c922fa2b3da0b605814facec391f4c365cbc359e +size 195123 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9jDfLBPWwZM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9jDfLBPWwZM.flac new file mode 100644 index 0000000000000000000000000000000000000000..fbe36b24c3d7b83a1068b0d3a50300181fb1db29 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9jDfLBPWwZM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ddba80c05bbc58238c4f944034d2830bea735f523780732729ae79d6b883cba +size 229475 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9lO8Hii2Q1I.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9lO8Hii2Q1I.flac new file mode 100644 index 0000000000000000000000000000000000000000..164660da4b9765868cb43eda3e2072b6201668f0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9lO8Hii2Q1I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a673ce4b1d0af63bf7f30090a578299c021871940e1cf6b6e98a14cf94badf6 +size 240465 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9o3monl8eBU.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9o3monl8eBU.flac new file mode 100644 index 0000000000000000000000000000000000000000..427154b56bb7c2e4b0a3d804479fd632633bfe8f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9o3monl8eBU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:129f6e2042f1076a2b1979982e9ae4ce5aca7df56af2b4d31efab991b1676152 +size 175613 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9qj7l2LE2PM.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9qj7l2LE2PM.flac new file mode 100644 index 0000000000000000000000000000000000000000..d7b8e8be22bbd8545cc062018a740b482be2caea --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9qj7l2LE2PM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a17474b35524b5fd6e352ecf81c8c32f6e4d064157b88b62a78b08a09aeb21c +size 231275 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9qwuWhM2C8Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9qwuWhM2C8Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..f73efddd82ee1efa321a63abc606fa1083fdee74 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9qwuWhM2C8Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3298f65b4ba01029799a8be8d3c3774e2bb451dbba914c3be7053bcb9c552af2 +size 198516 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9ucfsaBXB2Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9ucfsaBXB2Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..26606372c1c0d8f76cf44bdce6f0acc57ae1db58 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9ucfsaBXB2Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b37177835b147052917da056a8117cdd64e2a5c0c916950a00191e001fd5bb17 +size 170708 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9ucxDI35nig.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9ucxDI35nig.flac new file mode 100644 index 0000000000000000000000000000000000000000..451ae2330fc1054d45a6c15416359e65b8ae2231 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9ucxDI35nig.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:173052406bf8948e1d3fe558b4e3f7fd0afddd0c7e64f0eca45f5a68155cfb2a +size 267172 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y9wGJydlJvrk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y9wGJydlJvrk.flac new file mode 100644 index 0000000000000000000000000000000000000000..4986ef035e6bf10bb86c0234373bd1a6622c9e48 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y9wGJydlJvrk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f3433f858bac35c8dddc13d9c456e6642c19805e8d0b94975f585acc001f3227 +size 221433 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YA-7kQF3QN54.flac b/Foundation/Audio_Grounding_AudioGrounding/YA-7kQF3QN54.flac new file mode 100644 index 0000000000000000000000000000000000000000..bd0d95ea7f2aa7858a09939f968ac70b2ee56a5b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YA-7kQF3QN54.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2be06dcc478d5bb3fccbb28484c0e8854ec15e6623e01baf35fa32d4b8f1444 +size 135140 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YA02FI-n7uDk.flac b/Foundation/Audio_Grounding_AudioGrounding/YA02FI-n7uDk.flac new file mode 100644 index 0000000000000000000000000000000000000000..0c3a9ad08b49b5e03a67c1ef689489020d71e5bd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YA02FI-n7uDk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:466809dae83ac5f850ad624526925af97b1ee9f89858cd48ac94581b73e94397 +size 300772 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YA1iwaCZDWE4.flac b/Foundation/Audio_Grounding_AudioGrounding/YA1iwaCZDWE4.flac new file mode 100644 index 0000000000000000000000000000000000000000..3ed71777f21a94e88fc56239d3f5f2e72efb90de --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YA1iwaCZDWE4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fffb349f8f75b20668552cdb807e2fb2f6505d7ea645fb7c104e0492633dbee2 +size 232275 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YA21eA2M6cuM.flac b/Foundation/Audio_Grounding_AudioGrounding/YA21eA2M6cuM.flac new file mode 100644 index 0000000000000000000000000000000000000000..6600cf3ae76ec9d644c9566f17d97de8efe21fc4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YA21eA2M6cuM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa7d8eecc5ec353bc072d9a50e79144dd9b8f04ed1268f635d7d1e8b496f99c1 +size 226835 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAfrV3HG7JFw.flac b/Foundation/Audio_Grounding_AudioGrounding/YAfrV3HG7JFw.flac new file mode 100644 index 0000000000000000000000000000000000000000..857698258d90ae4d79b78f4cfeaaf77b2f5cc78b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAfrV3HG7JFw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6d6706086a25069371fdb9b774e2c15f702959e9a1cbb756350d0e73124b43ed +size 222381 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAgG_nhUurY4.flac b/Foundation/Audio_Grounding_AudioGrounding/YAgG_nhUurY4.flac new file mode 100644 index 0000000000000000000000000000000000000000..61b59a91496c968d80d6605e512752c5c6a16484 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAgG_nhUurY4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3489d0b63e990e331422b2bc9c98448ec560252bcb92d42521972dd54ae20b03 +size 255894 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAlPTUAK6ero.flac b/Foundation/Audio_Grounding_AudioGrounding/YAlPTUAK6ero.flac new file mode 100644 index 0000000000000000000000000000000000000000..d21d1af5f7bef8a367f0a9dc5b85d2c29acabf16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAlPTUAK6ero.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9db57892a65aaa238424045d718f58d96eb44ffb5e5a3130fb5c7110454ad3a +size 207209 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAlzK5RtVIyc.flac b/Foundation/Audio_Grounding_AudioGrounding/YAlzK5RtVIyc.flac new file mode 100644 index 0000000000000000000000000000000000000000..34691e09addf160b8c354ada5d4dc9efacee2ae2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAlzK5RtVIyc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9709f064fc0d3f6df457e29c1d13e557429da8b99f5143442ef6f51c2a1d3515 +size 276932 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAp8Eb8blOQA.flac b/Foundation/Audio_Grounding_AudioGrounding/YAp8Eb8blOQA.flac new file mode 100644 index 0000000000000000000000000000000000000000..6a928024cb8f969243c572c5bef9c850fda963a9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAp8Eb8blOQA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:318c4387fb9ab5924ac4e88208d5304f4f8ecfdafe0cfa43e17ac7988dca5400 +size 244077 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAtmv7HYs-uk.flac b/Foundation/Audio_Grounding_AudioGrounding/YAtmv7HYs-uk.flac new file mode 100644 index 0000000000000000000000000000000000000000..4dd0bb588d0375dedbb26bc3fd20a62207ebcf0a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAtmv7HYs-uk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:726d1b5326c804e64737f9213994268b2682eefb42cd1ea463dcd0888318a272 +size 232631 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YAw89ZmdNpYU.flac b/Foundation/Audio_Grounding_AudioGrounding/YAw89ZmdNpYU.flac new file mode 100644 index 0000000000000000000000000000000000000000..be5f36f2f5b5e50410b5e85a8318845bc5bba4a8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YAw89ZmdNpYU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdc4487258fe7ccec50ca6a5165bf2edc4b2a4fd45434b436a15308a7292a74b +size 210205 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YB10u_JR3sdQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YB10u_JR3sdQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..616717b63745f1cbaae1f77d301c7f43c194b0ef --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YB10u_JR3sdQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c786ffdea02569e96387e9da6e52d84c4e0803082deba61e0318eb58021481d7 +size 240550 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YB35iQ9_XLhk.flac b/Foundation/Audio_Grounding_AudioGrounding/YB35iQ9_XLhk.flac new file mode 100644 index 0000000000000000000000000000000000000000..c36c1b264583f8f2b222703e8de19713a308de51 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YB35iQ9_XLhk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:852da6beb45f02d138aafd574543812b88b03fdf61c167264d4b4c45ecd3591a +size 287115 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YB4FuuAYbE60.flac b/Foundation/Audio_Grounding_AudioGrounding/YB4FuuAYbE60.flac new file mode 100644 index 0000000000000000000000000000000000000000..31625f1817a4f14481fdfca83d99976776e7fab0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YB4FuuAYbE60.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d5dd20bbeed6924dc722c01deceb33d2df1af50dbfa2be65b110cd66e5fc7c48 +size 178106 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YB8VhLKlh9pU.flac b/Foundation/Audio_Grounding_AudioGrounding/YB8VhLKlh9pU.flac new file mode 100644 index 0000000000000000000000000000000000000000..979aeefcf7f3660eabf5653d0219a9ba82bd3b89 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YB8VhLKlh9pU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:92b2f47256241ca7d64f058471bb645c2a00b708ead0c81b2ea4fb9736833598 +size 132219 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBA-lFjpzad4.flac b/Foundation/Audio_Grounding_AudioGrounding/YBA-lFjpzad4.flac new file mode 100644 index 0000000000000000000000000000000000000000..a42f3d5b120563731d3f88cbc107b2943794b660 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBA-lFjpzad4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a5880921d48544bdcc58ed6f6503f31b16068418efdfbf0d5d98d5f41d758df +size 270373 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBBm27mZtcng.flac b/Foundation/Audio_Grounding_AudioGrounding/YBBm27mZtcng.flac new file mode 100644 index 0000000000000000000000000000000000000000..a6d18fd222bed4465b6bf972fdb4ed3037d45a08 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBBm27mZtcng.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77f030b22c5405e961157583d165a8429b133d088ff06f6a6c14cee21ebc4aac +size 270657 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBCrS7bCWPXs.flac b/Foundation/Audio_Grounding_AudioGrounding/YBCrS7bCWPXs.flac new file mode 100644 index 0000000000000000000000000000000000000000..d9f70b3722392aa01fefac95ca5e27c380ecb1f2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBCrS7bCWPXs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97e523f1fe71ecf6fc99aa6a481f3c1ef5df049f32e33479ae7e0b2397aa4520 +size 245142 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBCxuARguSBg.flac b/Foundation/Audio_Grounding_AudioGrounding/YBCxuARguSBg.flac new file mode 100644 index 0000000000000000000000000000000000000000..101bf66550cb579f439e9d520abf775777f8a3c3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBCxuARguSBg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b37726ceecab0d9519aecbc08e8ceb5e761affc5b2915d8c0e3fc5d4d52f33cf +size 287474 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBDXGM5qwT_w.flac b/Foundation/Audio_Grounding_AudioGrounding/YBDXGM5qwT_w.flac new file mode 100644 index 0000000000000000000000000000000000000000..644831ba2170865a90f14167a1d2d7db0e602d85 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBDXGM5qwT_w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60254d167affb297bda06311ea11626757c6229ec3832896f0d15325f82d97a3 +size 190426 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBFHGv_6hU0Y.flac b/Foundation/Audio_Grounding_AudioGrounding/YBFHGv_6hU0Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..cd4ca6a5271d00095ccf6325466dd2736d7e8216 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBFHGv_6hU0Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa2a28d637b8db124fdbd04cbc7b42eb2e2762896867e9784f94269fbb024aa1 +size 278598 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBFgEYIWrooc.flac b/Foundation/Audio_Grounding_AudioGrounding/YBFgEYIWrooc.flac new file mode 100644 index 0000000000000000000000000000000000000000..78a55e9ed02e9db6fddb4ad55bf0093487f3b4e5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBFgEYIWrooc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bf05eb7ccfd1a520a178905f6a590b58e6c71c30229f27f878569e65c09ccc57 +size 243063 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBHdi8AH1McE.flac b/Foundation/Audio_Grounding_AudioGrounding/YBHdi8AH1McE.flac new file mode 100644 index 0000000000000000000000000000000000000000..bf5a4a20d71f210ea64971d1e479731dee831fa7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBHdi8AH1McE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4a9e61dc759f5eceec39b1c1755b24c62da4b5caa77cc8ca61a20aa49d31d290 +size 216764 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBLTscl4P0RI.flac b/Foundation/Audio_Grounding_AudioGrounding/YBLTscl4P0RI.flac new file mode 100644 index 0000000000000000000000000000000000000000..75344ef27cdeb39c3dbb6af22dcbcc0aedeb9c4c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBLTscl4P0RI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b65b85af741dea4a1a41a6bd118b9b895650b9dcbac23afc882a44d889513850 +size 262381 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBLi6DZmr_fU.flac b/Foundation/Audio_Grounding_AudioGrounding/YBLi6DZmr_fU.flac new file mode 100644 index 0000000000000000000000000000000000000000..7fc9e08c929477abaaa8520a1908a8ebd362fb92 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBLi6DZmr_fU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:094e1b4c1c8d841d6928d46dcb7152fae3a1cb3e18aaca6ea667346b4092d4a0 +size 183245 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBML3g2YRXt0.flac b/Foundation/Audio_Grounding_AudioGrounding/YBML3g2YRXt0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c19272a602b2348728e38dd7db37b37d34936f86 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBML3g2YRXt0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0c6cf764a4013fc548927c56e061918b794be2ed91e2a19b5832af1f325589e +size 204679 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBXxlqaDvdaA.flac b/Foundation/Audio_Grounding_AudioGrounding/YBXxlqaDvdaA.flac new file mode 100644 index 0000000000000000000000000000000000000000..23fd2e7ec4d15da379574d8cd08d1d8b34742785 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBXxlqaDvdaA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb98090a5813c12c8107434d18a21a7fa9b518e0d9b01097c129e0a4ae569f7a +size 212612 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YBbbCArC3CC0.flac b/Foundation/Audio_Grounding_AudioGrounding/YBbbCArC3CC0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c397648931cbbf850207b2f16934467a0ab15ab5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YBbbCArC3CC0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59f4ec621665e3b0abcde6c487960b2471b0f0e585f5850adece15c7e8b2cc1e +size 203364 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YC0AY0FEhOxU.flac b/Foundation/Audio_Grounding_AudioGrounding/YC0AY0FEhOxU.flac new file mode 100644 index 0000000000000000000000000000000000000000..dcbfd00e7fd20375bea7220b653edf16f53c4f05 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YC0AY0FEhOxU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e25059e5569fa64b3838928f68a27d5e33b589aff20d884e0789092c9d623f38 +size 130237 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YC7yZovKH0Jg.flac b/Foundation/Audio_Grounding_AudioGrounding/YC7yZovKH0Jg.flac new file mode 100644 index 0000000000000000000000000000000000000000..49c54b9765ddfc1dec5f88b48462cdc515562862 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YC7yZovKH0Jg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:866c2c96f33ea760e308f98f2c32ae0d3ea8700f3dbacd9d1fe55ddc95cc9fc4 +size 256373 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YCbe2B6ohBpw.flac b/Foundation/Audio_Grounding_AudioGrounding/YCbe2B6ohBpw.flac new file mode 100644 index 0000000000000000000000000000000000000000..89b54080c92e5a25dfddd6f790432a4e05037ac8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YCbe2B6ohBpw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:651f5e882825e9cd323d5533627c74db09cc089db45f9e1519981cf3f1c08558 +size 181237 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YCh0LMmhBUg4.flac b/Foundation/Audio_Grounding_AudioGrounding/YCh0LMmhBUg4.flac new file mode 100644 index 0000000000000000000000000000000000000000..807ed6049482c1fbd49d28a85b09d85bcd7351c9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YCh0LMmhBUg4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e379b9d53ec65063b61a7915db58a5a35f121f1d915ccd5294b97f66ce1f353e +size 216014 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YChNWetbt6Xo.flac b/Foundation/Audio_Grounding_AudioGrounding/YChNWetbt6Xo.flac new file mode 100644 index 0000000000000000000000000000000000000000..fdd2035c37a2cebff070bc3bbd73d0a32f5c041e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YChNWetbt6Xo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7184e72dd5703b7d6e64402350ff3a0b0c36cdd3a3fb638954e6efc6cbb60e46 +size 251109 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YD1NopL9cFPQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YD1NopL9cFPQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..ab4efd3758c08af1c1f36dbd3fb9f0c39982cdbf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YD1NopL9cFPQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30a335ba35ef8e82cc9ab090ab78d58d2c17d5893cd3a690c18f57d95ae58ffd +size 263858 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YD2Xc_jZllDY.flac b/Foundation/Audio_Grounding_AudioGrounding/YD2Xc_jZllDY.flac new file mode 100644 index 0000000000000000000000000000000000000000..eedbb9fff342d79f85209a4bfeddaad2fa6e7375 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YD2Xc_jZllDY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1a2d53f25cd9a78cc4b03996046087078e3319a7423979cfb9f9cdd88d574f9 +size 165942 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YD6Xt-fOZy2k.flac b/Foundation/Audio_Grounding_AudioGrounding/YD6Xt-fOZy2k.flac new file mode 100644 index 0000000000000000000000000000000000000000..789ad31490ae3c41b70c1883dab2253b7bdadbf3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YD6Xt-fOZy2k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ed6a241b3ee89b7dc10ac5709b72ec1a3c237a21589c0d3448b30b3345cea4a +size 289783 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDMtUvX8y87M.flac b/Foundation/Audio_Grounding_AudioGrounding/YDMtUvX8y87M.flac new file mode 100644 index 0000000000000000000000000000000000000000..a1fb942c311b3af4d3ceeae6af470c8812a701ee --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDMtUvX8y87M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1aebc6a1430a145bb9c9dbcab51fca38e7524273a5caf65065a52ae35df67efa +size 162356 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDO_ZDe2CN5I.flac b/Foundation/Audio_Grounding_AudioGrounding/YDO_ZDe2CN5I.flac new file mode 100644 index 0000000000000000000000000000000000000000..224db31811bd7aff561e94563756ff72edce634d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDO_ZDe2CN5I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dc4fa4be5150f2f8954a074b364c8d972d2f190852aaf67c22a9abb840de5ae +size 193985 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDRSiR-dkmU8.flac b/Foundation/Audio_Grounding_AudioGrounding/YDRSiR-dkmU8.flac new file mode 100644 index 0000000000000000000000000000000000000000..5e90fb1120550de7b932963ddef75a5eb2c1ac9b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDRSiR-dkmU8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f27937c7e706425b83197a0d0a8883018f73f5f3d369465f4bd120a297850e72 +size 188181 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDcGnX0xLRLM.flac b/Foundation/Audio_Grounding_AudioGrounding/YDcGnX0xLRLM.flac new file mode 100644 index 0000000000000000000000000000000000000000..2be51e907e7ae412bfc3a7fabd0d0db5af9b906b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDcGnX0xLRLM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3a1a589f2aef4a11229f4255b83920fe53485446b426c6da6e3505d8a02cf8e +size 177154 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDcM_wehXfeg.flac b/Foundation/Audio_Grounding_AudioGrounding/YDcM_wehXfeg.flac new file mode 100644 index 0000000000000000000000000000000000000000..25fd41943ca5e059f440590f0771745ef6fa4e32 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDcM_wehXfeg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a80683b19e1f5dcd6aeade98f7cde2158d45253c6067d6367a6efcaf22ea073 +size 243157 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YDx7q7vzVFGE.flac b/Foundation/Audio_Grounding_AudioGrounding/YDx7q7vzVFGE.flac new file mode 100644 index 0000000000000000000000000000000000000000..671a1f444e4cc6598eb2413bdf965f2577ed7dda --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YDx7q7vzVFGE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:98ec336156c83b3a2c8f50bd0c23fc6e881382ea126ac98469b569cc8b6b7d6c +size 234193 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YE3sNwZlVRv4.flac b/Foundation/Audio_Grounding_AudioGrounding/YE3sNwZlVRv4.flac new file mode 100644 index 0000000000000000000000000000000000000000..1cd3d2f5bece036e26edba2d5b429cf856bddafd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YE3sNwZlVRv4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe2ac1406dafcc89cfab71b1cba98ddf2de052b96675e187f61361bba71ab6d8 +size 229273 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YE70Rf1Aol3k.flac b/Foundation/Audio_Grounding_AudioGrounding/YE70Rf1Aol3k.flac new file mode 100644 index 0000000000000000000000000000000000000000..465af63632385d1f47757aef5c4e3cbbd016b536 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YE70Rf1Aol3k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2165fe947a75cd6fa02e8774237832e3d80a6c4b5bd03eb229f66ccf526544a +size 259612 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YECPfahcqU4k.flac b/Foundation/Audio_Grounding_AudioGrounding/YECPfahcqU4k.flac new file mode 100644 index 0000000000000000000000000000000000000000..f17aa5bd4b97197212ffca96c3edd17196d0ea2d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YECPfahcqU4k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1a6ed9cb45aa0e3d917b7cadd3b077c89fb7988bac95b2d6a5d38b42d79e22d +size 267339 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YEONfjVbpIuM.flac b/Foundation/Audio_Grounding_AudioGrounding/YEONfjVbpIuM.flac new file mode 100644 index 0000000000000000000000000000000000000000..6afcbbfe9d46fefb6d2424c0c1c284cf97f39203 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YEONfjVbpIuM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f4de370de96bfe2512402ce39cdfa5dddfb3724382dde0c9e52650e3504ce2a6 +size 242977 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YEQV_KGRWvgg.flac b/Foundation/Audio_Grounding_AudioGrounding/YEQV_KGRWvgg.flac new file mode 100644 index 0000000000000000000000000000000000000000..698bd0c06e5ddf14dde4603fd9f47f69958ff9c7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YEQV_KGRWvgg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b3694cc01e1b67c8dc60cc641b8b1295738847e0ca483b837fd712ffc41598 +size 301048 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YETXMll5dC8Q.flac b/Foundation/Audio_Grounding_AudioGrounding/YETXMll5dC8Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..20a538b2595ae26cacfcb9c70b55c2b0c508aabd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YETXMll5dC8Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9004ec2ea035a4660b982930f8a91b8d401e8458f3e06786335591e5bc85adec +size 232196 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YE_Tqm4Twvf0.flac b/Foundation/Audio_Grounding_AudioGrounding/YE_Tqm4Twvf0.flac new file mode 100644 index 0000000000000000000000000000000000000000..2c6c873ea6b4f0e2046150be5ed35343a518ef7a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YE_Tqm4Twvf0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b369e185e9ee9419b86333cc40fc151c940c6c71a7fdd6315ef7666bbe801d31 +size 291768 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YEoPFW-CgltQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YEoPFW-CgltQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..04b84ee22724e68be346efffa430be58b504f8a3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YEoPFW-CgltQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f03d116a8b5708e8317e6bf32fc5df32c1dc2b939cc6c3c052d1e64fb9460ec0 +size 215228 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YEyjStn4c9lE.flac b/Foundation/Audio_Grounding_AudioGrounding/YEyjStn4c9lE.flac new file mode 100644 index 0000000000000000000000000000000000000000..64b2e581f16d15257e309105ec558a9bb5c8b82c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YEyjStn4c9lE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c3a21182a5dd4b94ab01cf7f31b199be8272db78fc402ef5b246b8d1699b899 +size 268545 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YFZuQBXARPLQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YFZuQBXARPLQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..7de7f0d70168d1106f351b802af908fa84da495c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YFZuQBXARPLQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:27bcd376f0b4d6dfbbedb5a053b05f5675b195d89970febe47a2148bc5c2be50 +size 243809 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YFcD3ukcMbCc.flac b/Foundation/Audio_Grounding_AudioGrounding/YFcD3ukcMbCc.flac new file mode 100644 index 0000000000000000000000000000000000000000..9670bbd0d71aaad5d40be8c4c4197603b9e80fcf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YFcD3ukcMbCc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9203d2fb4235c9f6c243d6d855c526e93624134b350bbc917d5121c40f9c4996 +size 258915 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YFjtO-xQD_To.flac b/Foundation/Audio_Grounding_AudioGrounding/YFjtO-xQD_To.flac new file mode 100644 index 0000000000000000000000000000000000000000..9996311ada7731f33cbeb0b3182dd07bff441037 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YFjtO-xQD_To.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:205e37879b46bd938879a95182e02969b41f7237f4ef6ae91a20d007272516a7 +size 251616 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YFkyO6R1Frgw.flac b/Foundation/Audio_Grounding_AudioGrounding/YFkyO6R1Frgw.flac new file mode 100644 index 0000000000000000000000000000000000000000..fdadbf0373b7b0d6b03cbee9b74510050c7f3e77 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YFkyO6R1Frgw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fb51ed6fd41107b9ab51973ad432ce4e80836636db5fa82d1b23bbf04edf00c +size 162366 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YFm4oMQOP6Nc.flac b/Foundation/Audio_Grounding_AudioGrounding/YFm4oMQOP6Nc.flac new file mode 100644 index 0000000000000000000000000000000000000000..b43cc755088e951089878eb9aa172a839bf81af7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YFm4oMQOP6Nc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65cb50d3fcc6ff0dd417a59a91d447ea6a3961e3eec3e2b5d7f50bb5679007d1 +size 244887 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YG9cpl27TYw8.flac b/Foundation/Audio_Grounding_AudioGrounding/YG9cpl27TYw8.flac new file mode 100644 index 0000000000000000000000000000000000000000..b3d6d40bfdc3bc3aec05759f6b5fee4097790b0f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YG9cpl27TYw8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8d6aea2def090613f4104d43a67e62953baa2b204914fd20d2eeb228f21343a1 +size 225391 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGB-RPZWnk50.flac b/Foundation/Audio_Grounding_AudioGrounding/YGB-RPZWnk50.flac new file mode 100644 index 0000000000000000000000000000000000000000..a3a4e73b19a4fa293dc8a240598d5bfc7c444bb5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGB-RPZWnk50.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c538dfce0aa209db5772126279ac7b0883cd246cdd9b6640d1d9c912fc92bbc7 +size 189426 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGFQnh84kNwU.flac b/Foundation/Audio_Grounding_AudioGrounding/YGFQnh84kNwU.flac new file mode 100644 index 0000000000000000000000000000000000000000..37d7157b9e0fc1517345b59bd13d956795d9b054 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGFQnh84kNwU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d76cbb872abf0f5b85f593b3ae4a8f14577e58844293d469071b6e6c00918558 +size 220491 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGJbfVBLsuQA.flac b/Foundation/Audio_Grounding_AudioGrounding/YGJbfVBLsuQA.flac new file mode 100644 index 0000000000000000000000000000000000000000..4c148ea104487de7f8e6b9aae9c806c8410d8b18 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGJbfVBLsuQA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ee1ea6643a2bbcf21499fc601fbca810397cb8d9d57eb422708a99864631d68 +size 217318 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGLD6PgwF7Mg.flac b/Foundation/Audio_Grounding_AudioGrounding/YGLD6PgwF7Mg.flac new file mode 100644 index 0000000000000000000000000000000000000000..1deba9da0d52f58a738a32f0bace0b4050e5bb3e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGLD6PgwF7Mg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:682e16aa466d5e95e3f688085f7d221bc07e376e07fa36bba9016c73182ae3ec +size 187558 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGdtMcEQmO4c.flac b/Foundation/Audio_Grounding_AudioGrounding/YGdtMcEQmO4c.flac new file mode 100644 index 0000000000000000000000000000000000000000..a6a2096167a33937cad154c8e7ac772f21309855 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGdtMcEQmO4c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ba8b1445b502a633be967013cb27e105563e90146dc52db617849b35c1c5e41 +size 263390 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YGoKR2cbKdxA.flac b/Foundation/Audio_Grounding_AudioGrounding/YGoKR2cbKdxA.flac new file mode 100644 index 0000000000000000000000000000000000000000..33f22ddb02f173f9b677ec248075bfe24c000c75 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YGoKR2cbKdxA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c9d2cec808b206f4afd8735befe02e29aaa7c42433f772bf7b1cf69d0c0fbd1 +size 237731 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHAjoHORA-Lg.flac b/Foundation/Audio_Grounding_AudioGrounding/YHAjoHORA-Lg.flac new file mode 100644 index 0000000000000000000000000000000000000000..5b9618defbc978944d0c14fccdbbefd04c870e47 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHAjoHORA-Lg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02157259fa1e6f125b1ceab8d450a19b9e76e8192ef424dd0e764c40d018ef3b +size 275136 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHEg1bmaRqlA.flac b/Foundation/Audio_Grounding_AudioGrounding/YHEg1bmaRqlA.flac new file mode 100644 index 0000000000000000000000000000000000000000..753f0cb973e14c4e3611df0fd1c1dd41d72aeec8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHEg1bmaRqlA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecaddd3461cffa43ac7414168ecb153749ae44f9d60a99f1a1b6d50886daf074 +size 272976 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHMoZBbpPBFk.flac b/Foundation/Audio_Grounding_AudioGrounding/YHMoZBbpPBFk.flac new file mode 100644 index 0000000000000000000000000000000000000000..1a884aea492e5a154e2cdd5cd8f792888b4d4760 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHMoZBbpPBFk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cd50d5fe1512f62de6c2879f4689211ff9ac5ad46a9b3e492b6c2fd389f85c67 +size 188692 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHWBzPVfLFHg.flac b/Foundation/Audio_Grounding_AudioGrounding/YHWBzPVfLFHg.flac new file mode 100644 index 0000000000000000000000000000000000000000..c4c1e08fb716890b5d8d4f376a646a85abf0fdec --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHWBzPVfLFHg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a72f90f4e80f9a7413ba5e65b97d7ce4a7f22144c76ac936fc10448a727dbd4 +size 219149 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHhwFPF8BwSk.flac b/Foundation/Audio_Grounding_AudioGrounding/YHhwFPF8BwSk.flac new file mode 100644 index 0000000000000000000000000000000000000000..6493b32a7bb20cebed75018864e01de15268476e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHhwFPF8BwSk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15907495bf4eeba010705e782a0671fef5f4f4f4cc9c12edb5925e759e2a7d98 +size 256934 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHmBBO3sV8DI.flac b/Foundation/Audio_Grounding_AudioGrounding/YHmBBO3sV8DI.flac new file mode 100644 index 0000000000000000000000000000000000000000..f9384ae4a55827d196ac904004f09aeb494002a9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHmBBO3sV8DI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed16c88b2cd02721c6eba0438122084a6ddd9b925f12741f360472009abbe77e +size 255771 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHp_uadxXh4g.flac b/Foundation/Audio_Grounding_AudioGrounding/YHp_uadxXh4g.flac new file mode 100644 index 0000000000000000000000000000000000000000..fa9bd6a832b3cfb7f5eb526bdead3ac21d69e84c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHp_uadxXh4g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d0478a6963077e49b4ce05bc4e985705b425ed3284f3203664380b15328cc6a6 +size 168385 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YHqnSyliKTKA.flac b/Foundation/Audio_Grounding_AudioGrounding/YHqnSyliKTKA.flac new file mode 100644 index 0000000000000000000000000000000000000000..09b3675267d7b71819f67e00d5466b6a76f0b334 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YHqnSyliKTKA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:19ee20b3b2f0b57bd0f5607478dc44c3c668361406f3a1cd2f05ea8d1563e0a1 +size 247155 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YI-btJhP_ki8.flac b/Foundation/Audio_Grounding_AudioGrounding/YI-btJhP_ki8.flac new file mode 100644 index 0000000000000000000000000000000000000000..a82ed87c1c4c006a2a8f057eeb676ea94798b974 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YI-btJhP_ki8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7da01423db72003a68bb5d004761b0320708d43ee152b9c8f07908245fc6c634 +size 234222 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YIPfaRF76gVU.flac b/Foundation/Audio_Grounding_AudioGrounding/YIPfaRF76gVU.flac new file mode 100644 index 0000000000000000000000000000000000000000..f5cb9cda47ffa7cc1d9de7021a16c7744a5523b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YIPfaRF76gVU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59314b944ac9bc8ee15aca03668c245051336451786ae1c5b965aa67c4a49df7 +size 267011 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YIZffL8aQW9w.flac b/Foundation/Audio_Grounding_AudioGrounding/YIZffL8aQW9w.flac new file mode 100644 index 0000000000000000000000000000000000000000..62c361921c74955b68d96ba8a4703722b0b53fd4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YIZffL8aQW9w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10eb82f4cb45f035fcc15d322059a576778c65b02e36cb0e58fdeaad21038f5d +size 211039 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YIclFv7GGoDI.flac b/Foundation/Audio_Grounding_AudioGrounding/YIclFv7GGoDI.flac new file mode 100644 index 0000000000000000000000000000000000000000..f0835861819ac0b97b1ffdd53f43a4ba27ff8402 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YIclFv7GGoDI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7b8114ce2e0c1e054f1b909633c5f1c0d67a831f766569dccdbd34b191147b0 +size 261069 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YIhjS6l-BTio.flac b/Foundation/Audio_Grounding_AudioGrounding/YIhjS6l-BTio.flac new file mode 100644 index 0000000000000000000000000000000000000000..cbc1d2730ddea2625fed9402e21551f8ef638462 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YIhjS6l-BTio.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac1abc046b296a26ed5936633921c4b9d874faede9db9bb9f1a03fbd10231ce8 +size 217266 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YIvBTc8y7s0M.flac b/Foundation/Audio_Grounding_AudioGrounding/YIvBTc8y7s0M.flac new file mode 100644 index 0000000000000000000000000000000000000000..66267fd9487311101e1ec9f24d4ccf2d9e35ed13 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YIvBTc8y7s0M.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d017de46621da466c8c7a5d219fcb8d82ac5b99ed9d87730d5bb3436c652127f +size 259710 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJ8Npt2M7tzE.flac b/Foundation/Audio_Grounding_AudioGrounding/YJ8Npt2M7tzE.flac new file mode 100644 index 0000000000000000000000000000000000000000..9d11c0a084316b2a68788933c9dae073c1816707 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJ8Npt2M7tzE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:504b5198b0631389946e506c811deeb9eec13eac815f8e2ea80415dd54429608 +size 87184 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJC8TDJnubFg.flac b/Foundation/Audio_Grounding_AudioGrounding/YJC8TDJnubFg.flac new file mode 100644 index 0000000000000000000000000000000000000000..b0aa5b45c5e6c64070d4005ddc11097c6335f62e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJC8TDJnubFg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e376493fe568cb84b2bcddd9bf820c78fbc24143bb0dbb37acd12aa26eb3ded3 +size 176110 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJFk4Qyn58CY.flac b/Foundation/Audio_Grounding_AudioGrounding/YJFk4Qyn58CY.flac new file mode 100644 index 0000000000000000000000000000000000000000..0e17bc437b242477ec5331e7662b0975c193e962 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJFk4Qyn58CY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1577fb5efdbccf16c821e1fc9b09ae738b952a5b3d8831bba98f60173730aee4 +size 241155 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJP_czEZIOb4.flac b/Foundation/Audio_Grounding_AudioGrounding/YJP_czEZIOb4.flac new file mode 100644 index 0000000000000000000000000000000000000000..843c159c139428fa0504829e5eae50fd5d69ddc3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJP_czEZIOb4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fbe1bf9c05b2f45f7b06a3a7f0e151d88369cfdd48ae395df0a43bbd5707dcc0 +size 101892 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJcT3vhg_Ie8.flac b/Foundation/Audio_Grounding_AudioGrounding/YJcT3vhg_Ie8.flac new file mode 100644 index 0000000000000000000000000000000000000000..9716c14b9eabd6e3c3e22aa7d61072bbfe1b2032 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJcT3vhg_Ie8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:326fe2c34428eb0ec5c2c09a09402842400d3cbd80f9f39476998da32f5713bf +size 254173 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJiPwWia5ryk.flac b/Foundation/Audio_Grounding_AudioGrounding/YJiPwWia5ryk.flac new file mode 100644 index 0000000000000000000000000000000000000000..57ee40e8288d4586d049f3cffa6e269895b59a01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJiPwWia5ryk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e0b026983c3750d41439cf6580556318a0eaad250d4263b91905dbe05b8ea777 +size 265644 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YJwl2o6jPXkc.flac b/Foundation/Audio_Grounding_AudioGrounding/YJwl2o6jPXkc.flac new file mode 100644 index 0000000000000000000000000000000000000000..de2c93b08a79e0a374ad3786f67d2d12c7ffad8d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YJwl2o6jPXkc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cec5c5642e598a94f43cd34a8d44f23970eb3648504753731066782a514a343b +size 246353 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YK6-Xm6FnX8s.flac b/Foundation/Audio_Grounding_AudioGrounding/YK6-Xm6FnX8s.flac new file mode 100644 index 0000000000000000000000000000000000000000..6c27ed12b0de86f784fba5d6dcff3bcf23b2397e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YK6-Xm6FnX8s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eea5c1d51569562862e1f9a598ec88cacc168bc81a47e470786b790291249d10 +size 275834 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YK9lmiWhXu1s.flac b/Foundation/Audio_Grounding_AudioGrounding/YK9lmiWhXu1s.flac new file mode 100644 index 0000000000000000000000000000000000000000..31b904e2ab4d2e64b3b32beb521a8c5acb468a7d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YK9lmiWhXu1s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:abcdf2214d24837e54289a051484e5dc092963e4530ee1436d97dd90cc9059d1 +size 144025 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YKVbmN9ZRg5Q.flac b/Foundation/Audio_Grounding_AudioGrounding/YKVbmN9ZRg5Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..1efe410414036f8e93177f188c70de97694004fe --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YKVbmN9ZRg5Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:674480ca1ba844543edd2ff13f9889f31582c0a0e251d39aeb85dfec0e51e56b +size 243293 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YKniufpODcOg.flac b/Foundation/Audio_Grounding_AudioGrounding/YKniufpODcOg.flac new file mode 100644 index 0000000000000000000000000000000000000000..356876381dd1dce467c2b2bb80b0f5ff07d9c731 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YKniufpODcOg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bcc18ec55af72c97eb8719364d37d29d21f16c2d1b65bdfcbefcfbeeb4157cba +size 220676 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YKq0Dbp3C4d0.flac b/Foundation/Audio_Grounding_AudioGrounding/YKq0Dbp3C4d0.flac new file mode 100644 index 0000000000000000000000000000000000000000..c0b01384d9afd549eb75b08cf742e3279459b182 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YKq0Dbp3C4d0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e80a7d600c8060c70fa7a348974e62b8f4b312c642220ba3954396887f19cbc2 +size 200088 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLBixNsRsbpE.flac b/Foundation/Audio_Grounding_AudioGrounding/YLBixNsRsbpE.flac new file mode 100644 index 0000000000000000000000000000000000000000..78f10e9211e0c9aef23325b2c32497e8fc3cca9f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLBixNsRsbpE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0cda7a08c4796991b80c28495640a0d5aa2a4ee79fd1a1bac31ad320a05d4805 +size 204429 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLBpmR72vQBk.flac b/Foundation/Audio_Grounding_AudioGrounding/YLBpmR72vQBk.flac new file mode 100644 index 0000000000000000000000000000000000000000..d8d6c6879ff6bf9b1c04331ad252a5b438374373 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLBpmR72vQBk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6ceb88fe60de69ad01c3903b9bfb395e5a5e2e5dd637b6ed9274cbfb2b89948 +size 187010 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLbJZuD_Y_uM.flac b/Foundation/Audio_Grounding_AudioGrounding/YLbJZuD_Y_uM.flac new file mode 100644 index 0000000000000000000000000000000000000000..e489906c4ae7e4cb19320b586cb7601b9f555702 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLbJZuD_Y_uM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:15a248d382dd1d8a3c50f62bb1d6bea525e3a885df80cd17b902badf27218215 +size 211496 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLlffd6cn5jY.flac b/Foundation/Audio_Grounding_AudioGrounding/YLlffd6cn5jY.flac new file mode 100644 index 0000000000000000000000000000000000000000..45d324696cf38abd06c1c189c101a6cd05f8698a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLlffd6cn5jY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f56c4d7fb10ab51e12e0efb6886dc250d8114ed886b22c6aaf77804aa3f48784 +size 272398 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLyeZhAu7yh4.flac b/Foundation/Audio_Grounding_AudioGrounding/YLyeZhAu7yh4.flac new file mode 100644 index 0000000000000000000000000000000000000000..8e6fb1e0c2889131844a3b5e185cefdd81e38d57 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLyeZhAu7yh4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1736edb349c70de0a2ffdf4a5e380c73025bc36285c013d4e8c6d2a3e50fddac +size 172888 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YLzcNa3HvD7c.flac b/Foundation/Audio_Grounding_AudioGrounding/YLzcNa3HvD7c.flac new file mode 100644 index 0000000000000000000000000000000000000000..7bc5c8ad6804aaf207476362bf6550b8d4b79afa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YLzcNa3HvD7c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e60430ed04bc17b637a5ff3d8d75b9dd389b98c50e8dbd5d82667887c70102ee +size 242634 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YM7Ba3iTY0rE.flac b/Foundation/Audio_Grounding_AudioGrounding/YM7Ba3iTY0rE.flac new file mode 100644 index 0000000000000000000000000000000000000000..a87bda14b0c8919ba52277c4ffa69eae28ffe1bb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YM7Ba3iTY0rE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:067803d6824c349c94a00485f4da60a77b8ce24221f62c8d513aa86ab65bd9ef +size 254568 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YMM9O504q7Zc.flac b/Foundation/Audio_Grounding_AudioGrounding/YMM9O504q7Zc.flac new file mode 100644 index 0000000000000000000000000000000000000000..8684fc1e73a33e16304011ff752bdd0063193174 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YMM9O504q7Zc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b76050dc1da66081880a7310a19d36f570b0eeeb74530334f9e64a5244e1ece +size 203985 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YMbmDn7ahGEg.flac b/Foundation/Audio_Grounding_AudioGrounding/YMbmDn7ahGEg.flac new file mode 100644 index 0000000000000000000000000000000000000000..4eb96d53710adcb145510470995646322f4cc0f2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YMbmDn7ahGEg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52af105661091e9930714bac0f3122b1e9d0de31214ac7669f7972a6aa728a65 +size 181973 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YMlb1RW4-LRg.flac b/Foundation/Audio_Grounding_AudioGrounding/YMlb1RW4-LRg.flac new file mode 100644 index 0000000000000000000000000000000000000000..c91eb8ab299b2b7867a21a7a2ff0def93153fd16 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YMlb1RW4-LRg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c42836bd5368f335fece6fcfa5eac102f41e9e69898d69dc0efe0bad55f3af8 +size 251578 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YMtmE6qBxURA.flac b/Foundation/Audio_Grounding_AudioGrounding/YMtmE6qBxURA.flac new file mode 100644 index 0000000000000000000000000000000000000000..8658d84dde642c91f68102292725e8e7c4c4c5bf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YMtmE6qBxURA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66271118d5695a538d6c1dbe3061dd369d3d5834a96c543fd9df7a9d98d00184 +size 248595 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YMvHpNzDpC6Q.flac b/Foundation/Audio_Grounding_AudioGrounding/YMvHpNzDpC6Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..ca8445c0385590a4f6b99c7a57196b5e75d8973e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YMvHpNzDpC6Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a83b5dbcf7858251b8074097a2e6bfea4c827943ddc8d7207652493ce38eca59 +size 219420 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YN20hpbx31NM.flac b/Foundation/Audio_Grounding_AudioGrounding/YN20hpbx31NM.flac new file mode 100644 index 0000000000000000000000000000000000000000..5d080157fec55355b71315a85c219849e426a750 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YN20hpbx31NM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:212829b66d65d865740601d6b1f5fa2629c5b0491d156d44797e8c7c19fc89eb +size 193706 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNDaVSIJaXVs.flac b/Foundation/Audio_Grounding_AudioGrounding/YNDaVSIJaXVs.flac new file mode 100644 index 0000000000000000000000000000000000000000..dfb86f02be6b0a36d8cec6dd61767125ebf38286 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNDaVSIJaXVs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b5d6def008eb9fd2ed0b1fcd1ba15672777b0524a756db82af5b88293aefd19 +size 251018 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNGPY84KXaAY.flac b/Foundation/Audio_Grounding_AudioGrounding/YNGPY84KXaAY.flac new file mode 100644 index 0000000000000000000000000000000000000000..66f3d203c6a747a7a68ac5a0187ab862a509ca2d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNGPY84KXaAY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:086b5c9d3220df6a66a64bc535ac41eeaa33ace59cf56fff0e255a4100618cd8 +size 167458 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNUCh9Xtbjkw.flac b/Foundation/Audio_Grounding_AudioGrounding/YNUCh9Xtbjkw.flac new file mode 100644 index 0000000000000000000000000000000000000000..ad159900ea2f16f1f5bab7e3479a7623e3043c2a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNUCh9Xtbjkw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ad04e84562929b43b53f5570c7740514372d1cea9aa6102a96723a1ff4e23f0 +size 254047 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNZMDi4ZRjjQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YNZMDi4ZRjjQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..cd0f39e362601a88b7b0d615e054a25cd7ba6f14 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNZMDi4ZRjjQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7adc75ebbf58bfc834756bd4fade50316056789cb6466925844cd7b577c436bf +size 216027 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNeWW30WZjPc.flac b/Foundation/Audio_Grounding_AudioGrounding/YNeWW30WZjPc.flac new file mode 100644 index 0000000000000000000000000000000000000000..f55f378ccfaa77391ee92f5bec30162bcaa22239 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNeWW30WZjPc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41d776e9532203891fb56b8e3e57744f8bc64006ed31c6c654300918c111009b +size 193578 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNek57Uq--bw.flac b/Foundation/Audio_Grounding_AudioGrounding/YNek57Uq--bw.flac new file mode 100644 index 0000000000000000000000000000000000000000..b1dfbf3256624988781559b85b55ba2959550722 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNek57Uq--bw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2df7ae6fd6da1cf473925f4538fca1ceb82b7bd742164b92f69f4448d65fa932 +size 201275 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNeuFvAzcDaI.flac b/Foundation/Audio_Grounding_AudioGrounding/YNeuFvAzcDaI.flac new file mode 100644 index 0000000000000000000000000000000000000000..92a2cceb7ede00b461a9e4a99c1c81574af4f830 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNeuFvAzcDaI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bceb29786417cabf06de4598b25d5ea8578dfff9f9284c5f87c6509cca6ff0d2 +size 254129 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNfGOKIniyyE.flac b/Foundation/Audio_Grounding_AudioGrounding/YNfGOKIniyyE.flac new file mode 100644 index 0000000000000000000000000000000000000000..1ba194bf4a48c4644c70733054d2ec23129b8b6d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNfGOKIniyyE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0bf26e34cfe33c5aaf0137d0c0cea1f1855f80c77e8eefdc896a83c42f60bb8 +size 202334 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YNnXucrfqnJ8.flac b/Foundation/Audio_Grounding_AudioGrounding/YNnXucrfqnJ8.flac new file mode 100644 index 0000000000000000000000000000000000000000..1ecc10afcd9d35a304e8e4a58faa2ebb6bebb044 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YNnXucrfqnJ8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ce917bec69eaa8d1faf51ccf3d22ae1e22108515fb4ed2ed4b5d9aa73e52f66 +size 239941 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YO5EGxn49f5w.flac b/Foundation/Audio_Grounding_AudioGrounding/YO5EGxn49f5w.flac new file mode 100644 index 0000000000000000000000000000000000000000..0a30a9222a1f8080f170ce1a0298f7a61fd5d643 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YO5EGxn49f5w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e185f23fed83b2432e34b1044d2ad8194f6a8412bf69a974819262c91a5192e9 +size 208268 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YO5wKqu7fcyw.flac b/Foundation/Audio_Grounding_AudioGrounding/YO5wKqu7fcyw.flac new file mode 100644 index 0000000000000000000000000000000000000000..202b30003f77f7f700abc3533a72e0e952a6a8be --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YO5wKqu7fcyw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6e35b1ebc30db148b7907999337f081d71d2e3a2d1ed01ef885742585bf93c8 +size 277588 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YO9xYzACrBjA.flac b/Foundation/Audio_Grounding_AudioGrounding/YO9xYzACrBjA.flac new file mode 100644 index 0000000000000000000000000000000000000000..22561816700dc0c4c640b74705653499ce845a0e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YO9xYzACrBjA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2b883b39db42d5a83500c98dd786532c1ce4aae3d36a0a87302b6cacb973c21f +size 286995 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YOVS0Z30_mMU.flac b/Foundation/Audio_Grounding_AudioGrounding/YOVS0Z30_mMU.flac new file mode 100644 index 0000000000000000000000000000000000000000..a9a5a38a98a6446e9c0b3968e1e8cd22dbd8c11c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YOVS0Z30_mMU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfa9f37539f744ead74da8c9571d6c7b13c983265a8f05bf0984a54f209e0212 +size 197131 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YOY5503vDeA8.flac b/Foundation/Audio_Grounding_AudioGrounding/YOY5503vDeA8.flac new file mode 100644 index 0000000000000000000000000000000000000000..a00f38eb570af1c5d61dadf516698232b69952b0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YOY5503vDeA8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:431fd736b8461417d8ca2b5757a51ac220ff88218a34b99ba3dc210509bbfba2 +size 224791 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YOaQR4KDY1Ug.flac b/Foundation/Audio_Grounding_AudioGrounding/YOaQR4KDY1Ug.flac new file mode 100644 index 0000000000000000000000000000000000000000..e89c2c50b1fb47e7ce96abe3728fa192f1af38bf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YOaQR4KDY1Ug.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c006eb1d9278d0e0633843bb4dcb3e00c5dc6a91dcc5224f658eb0c8481303ea +size 209522 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YOpmtLXu5NJI.flac b/Foundation/Audio_Grounding_AudioGrounding/YOpmtLXu5NJI.flac new file mode 100644 index 0000000000000000000000000000000000000000..2c681c43418599cdeb367dc593bd97f6f51ce0c1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YOpmtLXu5NJI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60730a4057c618ba3f81d2d19eae8d97686170bf93a6321fab7f01f121a49ee3 +size 255380 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YOt0bN_hz2ec.flac b/Foundation/Audio_Grounding_AudioGrounding/YOt0bN_hz2ec.flac new file mode 100644 index 0000000000000000000000000000000000000000..be5cc5b7afe0f1d174312c65a3b972522dab88c5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YOt0bN_hz2ec.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d655e353816cfbe3cc82ce9c9842a67bfd9dc188de1395bc4cb57858d4cd3f6b +size 269685 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YP808463iRRU.flac b/Foundation/Audio_Grounding_AudioGrounding/YP808463iRRU.flac new file mode 100644 index 0000000000000000000000000000000000000000..e8a6ce633e50eadc248635d1d863c6fe6a7ea564 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YP808463iRRU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:42af4dfed8451e22b41f70283e8380ea2ab4a3799c1ecfbb77b8b945879fd7c6 +size 152491 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YPRBAm39xujU.flac b/Foundation/Audio_Grounding_AudioGrounding/YPRBAm39xujU.flac new file mode 100644 index 0000000000000000000000000000000000000000..11428644e6d655a731012ee6e2564308896526fc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YPRBAm39xujU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0093c6608968e3ca0a8b49e22cf3ff04fd0ce8c02d991bbb95ee46fd7fa848bb +size 260098 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YPSdbflW9mQ0.flac b/Foundation/Audio_Grounding_AudioGrounding/YPSdbflW9mQ0.flac new file mode 100644 index 0000000000000000000000000000000000000000..18a74ba6b928d13acb7cede6a3e09cba3d226c57 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YPSdbflW9mQ0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:91dfb9ed9a8e80b0c144b4df15b9c367afc20a093fc9998376698adc74ff00bf +size 276605 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YPY_wYey3P14.flac b/Foundation/Audio_Grounding_AudioGrounding/YPY_wYey3P14.flac new file mode 100644 index 0000000000000000000000000000000000000000..3df4776539c8b20548df2f62b5abbd43a30b0a0d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YPY_wYey3P14.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65841c7e19881e5e94b901e42dd3651a7cee9cb04105fbd1e2eb6ea1a198d83b +size 274108 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQBLpskcb_As.flac b/Foundation/Audio_Grounding_AudioGrounding/YQBLpskcb_As.flac new file mode 100644 index 0000000000000000000000000000000000000000..caf5516912a6866df053e3063fe7e0b398070db3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQBLpskcb_As.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e1872d3f8f6bf5f0840d2ac7d925c37b02e198213db703ee072d110e78b14458 +size 261356 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQHVho4OnVtc.flac b/Foundation/Audio_Grounding_AudioGrounding/YQHVho4OnVtc.flac new file mode 100644 index 0000000000000000000000000000000000000000..d66cc99b52f631728300655c1d432d827fba680b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQHVho4OnVtc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69544ea5dd8b7d37a698c422ceb629476bc1bdbb2deedbf1a899b676ef4bf7b5 +size 249784 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQWVZ3oMc4g0.flac b/Foundation/Audio_Grounding_AudioGrounding/YQWVZ3oMc4g0.flac new file mode 100644 index 0000000000000000000000000000000000000000..aaf826cf2130c419e2b97347f94a2f004039cd07 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQWVZ3oMc4g0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1c1307244aa3cce8915ecf7f99e652e2cdea4d0eab220e00243ea6b77bbdd0bf +size 231800 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQWWAj4WZuyU.flac b/Foundation/Audio_Grounding_AudioGrounding/YQWWAj4WZuyU.flac new file mode 100644 index 0000000000000000000000000000000000000000..e45b0a706dd498506f171f57075eaf0dfb79671d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQWWAj4WZuyU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a8d4f16a94e2b3dae5e6844c93290515c64cc559b7df79e0adc505f5b8fb239 +size 201593 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQYQvPofLYOA.flac b/Foundation/Audio_Grounding_AudioGrounding/YQYQvPofLYOA.flac new file mode 100644 index 0000000000000000000000000000000000000000..977041dc627d3f66dfc96d0a289bbcc5130a8dd9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQYQvPofLYOA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d19d7c47aeaa92dd53ac84fd114507833867187491e75592ca776a68e2ea5394 +size 185591 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQa8QRHe31m8.flac b/Foundation/Audio_Grounding_AudioGrounding/YQa8QRHe31m8.flac new file mode 100644 index 0000000000000000000000000000000000000000..e51a6a4ccf8f4c36facbc688a1194d4d4dd420d5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQa8QRHe31m8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24547b39395e89a18c545e5954bb9360caea75165c166f6e1363d57304b51b4c +size 193775 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YQvATUKXYFBs.flac b/Foundation/Audio_Grounding_AudioGrounding/YQvATUKXYFBs.flac new file mode 100644 index 0000000000000000000000000000000000000000..e45f62d5bf3a5abbe08013dc7a0193faf7ef5f42 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YQvATUKXYFBs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:523aaff2a3ff7d20ecc3d0aa151172c1de33da42cef52f450171b179ff4ec7d6 +size 257284 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YR79kDKN1-to.flac b/Foundation/Audio_Grounding_AudioGrounding/YR79kDKN1-to.flac new file mode 100644 index 0000000000000000000000000000000000000000..7e24e452d9da8f04cd41fd174fdd44f13a943d9d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YR79kDKN1-to.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a1641887c5274b3a2893663c0e4bba1f96ef8b3616284443f62fb7da17f7210 +size 216847 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRGxGn-VGW5s.flac b/Foundation/Audio_Grounding_AudioGrounding/YRGxGn-VGW5s.flac new file mode 100644 index 0000000000000000000000000000000000000000..ca7a7429c6b504a24d689ebc8462b10337aed55d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRGxGn-VGW5s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:413a54eaec3b17d683d8261bed776b73d371ce6d34ddba222118eeb64b6fbc77 +size 289932 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRJrJAyUZwc4.flac b/Foundation/Audio_Grounding_AudioGrounding/YRJrJAyUZwc4.flac new file mode 100644 index 0000000000000000000000000000000000000000..a1ac7cb9609900f3b2b2da252e57fb2a84e13744 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRJrJAyUZwc4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:429ec9ebfa44b4d479b293c82fa6dd8dd6afb82d5e22d9527465114cb0420112 +size 227883 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRLYc5lfS3eE.flac b/Foundation/Audio_Grounding_AudioGrounding/YRLYc5lfS3eE.flac new file mode 100644 index 0000000000000000000000000000000000000000..0f1051aab122ebe6b5ec3d161969e8931157f6da --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRLYc5lfS3eE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e63d7b1e98909e51d8b3065f2e28e37fe6b7f096e9f90096e60840b7055c1044 +size 237110 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRPvqI4KYiO4.flac b/Foundation/Audio_Grounding_AudioGrounding/YRPvqI4KYiO4.flac new file mode 100644 index 0000000000000000000000000000000000000000..53c559472f26f9cf780aafb69f2025b623d67167 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRPvqI4KYiO4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8fc31670553d6e298c7a1cca1f588e5d8eaeaa8cfc775043c67401cdcdb7feb +size 165695 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRZZKDnECWyQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YRZZKDnECWyQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..797f00a9ce6fd8cd2c1d089b118fb641451ddc04 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRZZKDnECWyQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:689134bace782536042055f69ae0521162b5694c5797c7ec0ece66c8a29ab868 +size 257049 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRdHDbFchcV8.flac b/Foundation/Audio_Grounding_AudioGrounding/YRdHDbFchcV8.flac new file mode 100644 index 0000000000000000000000000000000000000000..a0740bb4fdfde8b622a5839abad6f70bb956b1ff --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRdHDbFchcV8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e4fe9ec118dbb7a0e02c2e139b36c9229196ede6d951068ddc4a5f66bc23ba30 +size 175078 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YRmXHbnsKe0g.flac b/Foundation/Audio_Grounding_AudioGrounding/YRmXHbnsKe0g.flac new file mode 100644 index 0000000000000000000000000000000000000000..1eeded21c93005e5ffb762787ab80086ab49a3f4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YRmXHbnsKe0g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4667659da449d59918bc007853b905dec85dd93b8a10de69ad3d28ac9bbb4bfd +size 277156 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YS3dsryb4eos.flac b/Foundation/Audio_Grounding_AudioGrounding/YS3dsryb4eos.flac new file mode 100644 index 0000000000000000000000000000000000000000..71d7f37b1ceea20e3dd1136a24e1b30982c0712a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YS3dsryb4eos.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:53d758d446ca9e63fcf851be87384b3e248587832045fe616e3518d658f1c818 +size 231944 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSEHB4eYLLPc.flac b/Foundation/Audio_Grounding_AudioGrounding/YSEHB4eYLLPc.flac new file mode 100644 index 0000000000000000000000000000000000000000..236c8bf055d3286aa72f8a86034b659450159906 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSEHB4eYLLPc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cbae42c34b108c54110f96b977cd2b86573570f6c0689239608daa78319938a +size 194869 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSEql80jbheU.flac b/Foundation/Audio_Grounding_AudioGrounding/YSEql80jbheU.flac new file mode 100644 index 0000000000000000000000000000000000000000..9da8d6d2d7debea5c78c4e5f7a7a991a26cbb596 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSEql80jbheU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b1fd21e580d057da83b37c6268160bf33163efada236819469a8efb3a287ea51 +size 264618 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSNIaYhri76w.flac b/Foundation/Audio_Grounding_AudioGrounding/YSNIaYhri76w.flac new file mode 100644 index 0000000000000000000000000000000000000000..a1f520da546efa03af583849060b7456f81a1852 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSNIaYhri76w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf082421d21b8440af26528f5c5d5cf1af0fd311ea62f138a586e26698e49b12 +size 227944 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSTIzm2JFbuc.flac b/Foundation/Audio_Grounding_AudioGrounding/YSTIzm2JFbuc.flac new file mode 100644 index 0000000000000000000000000000000000000000..78ee82f7a6e45c11b371aeaca551cd85628963eb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSTIzm2JFbuc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c9c044602625ca9b3cfeb50b036d9f2b8f13e154eda65fbc6cac062254fe911 +size 273775 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSiMabSrCdJw.flac b/Foundation/Audio_Grounding_AudioGrounding/YSiMabSrCdJw.flac new file mode 100644 index 0000000000000000000000000000000000000000..9e9a06b60449bdacd4e1b0fc699cdc6113a53c21 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSiMabSrCdJw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3342e26054c3e925c36c8224cb33e1725ae8b6b0b1adcd4a67eb7cea5e92408f +size 254393 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSmB8tjUGVNE.flac b/Foundation/Audio_Grounding_AudioGrounding/YSmB8tjUGVNE.flac new file mode 100644 index 0000000000000000000000000000000000000000..2b4fed032a7399ab8d501b0f99d10bb24c3fcf5e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSmB8tjUGVNE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f9a67ddca5b6caaadcd91390588812295e0788045f0562eb2c624245288577dc +size 189550 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSrobrl1Lh1A.flac b/Foundation/Audio_Grounding_AudioGrounding/YSrobrl1Lh1A.flac new file mode 100644 index 0000000000000000000000000000000000000000..68cd01e4c5e2ed4ea88680baa45a384df910a270 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSrobrl1Lh1A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e436d53b4b59d899307923920f13d0873df86fd99e35b99ef697b596de59d49a +size 218867 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YSz7po5sNgQw.flac b/Foundation/Audio_Grounding_AudioGrounding/YSz7po5sNgQw.flac new file mode 100644 index 0000000000000000000000000000000000000000..7ca6fa18b8a9b2a5be745f097b61d010c56998e5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YSz7po5sNgQw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad4a35d599a376da8c689677af7e692e4527562543ba961b6d9878fe0eae8860 +size 206859 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YUR-fBo0fmRo.flac b/Foundation/Audio_Grounding_AudioGrounding/YUR-fBo0fmRo.flac new file mode 100644 index 0000000000000000000000000000000000000000..d301d558bde512931d63b535e76e3b7c0f3304a6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YUR-fBo0fmRo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dec146b9a90f3be0684181f0bf2797fed20a0d5c873848090c577f3563d64293 +size 258808 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YUW8czMvlmK4.flac b/Foundation/Audio_Grounding_AudioGrounding/YUW8czMvlmK4.flac new file mode 100644 index 0000000000000000000000000000000000000000..a4574d5d08d34f3cb1866c0fa55e8845fdf75966 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YUW8czMvlmK4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fad05d67da1d19a8acdd043482044c6b77aa8881b8f48a19bea30da67d44618 +size 261476 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YUsncYzpYSk8.flac b/Foundation/Audio_Grounding_AudioGrounding/YUsncYzpYSk8.flac new file mode 100644 index 0000000000000000000000000000000000000000..5215d4dadf7728768666da1cfb849e1a53e85fcb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YUsncYzpYSk8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6aa28a635fb7f73918493bdd299dc5d33c9493c3950d84d525e096fc5288cb17 +size 230805 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YUwQhrCDchio.flac b/Foundation/Audio_Grounding_AudioGrounding/YUwQhrCDchio.flac new file mode 100644 index 0000000000000000000000000000000000000000..40693876f24bbac7c59a7cdea9216c407f1e4cab --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YUwQhrCDchio.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:54e3f6d5ea4013323fbdc426972f5c45dfe08a2437b9e7ff3d1449b33742ee27 +size 213181 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVOXl8iR-HnI.flac b/Foundation/Audio_Grounding_AudioGrounding/YVOXl8iR-HnI.flac new file mode 100644 index 0000000000000000000000000000000000000000..336050a90be0fa0edaf43e1143bb5b15da73886d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVOXl8iR-HnI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dffb716910fd3112578c67a5f899ebf3b65e82a6d8d6f52701e7fdb37ff67a26 +size 241450 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVb9krAYFfWE.flac b/Foundation/Audio_Grounding_AudioGrounding/YVb9krAYFfWE.flac new file mode 100644 index 0000000000000000000000000000000000000000..bb82c2650300e2cfec137ca1fb43c75a250597ce --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVb9krAYFfWE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32c8db351294b9f2f028bf978bfbe1656c2644ca85d00d97b6aa25b7f218b8e7 +size 243572 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVc5wX72RjN0.flac b/Foundation/Audio_Grounding_AudioGrounding/YVc5wX72RjN0.flac new file mode 100644 index 0000000000000000000000000000000000000000..742505716228f4600ed00ddf397898782644e9e4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVc5wX72RjN0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6737e8cc90f4af8aea156d437595059bc44a3409e89b59da7d951dea2d8b0234 +size 199516 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVhIYEoQseoM.flac b/Foundation/Audio_Grounding_AudioGrounding/YVhIYEoQseoM.flac new file mode 100644 index 0000000000000000000000000000000000000000..4463c11b775afb4deb2a3790c2ded46ee9e69e9f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVhIYEoQseoM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0ae8d2505c2bf3dea95f370e136056532d2e68ee9cd2c2abaf9ba2651181c3b0 +size 201411 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVmSHONdqvng.flac b/Foundation/Audio_Grounding_AudioGrounding/YVmSHONdqvng.flac new file mode 100644 index 0000000000000000000000000000000000000000..7de2202a84ddd2596f8bd8cac4e683def000450f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVmSHONdqvng.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b781350266950991f2e57cca0416aa04973075e3122bb2aeb46543abed47011 +size 260403 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YVrJlserGqxI.flac b/Foundation/Audio_Grounding_AudioGrounding/YVrJlserGqxI.flac new file mode 100644 index 0000000000000000000000000000000000000000..8065da641aee1783ea1c848d590d02fa3e019f01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YVrJlserGqxI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7479d7d5da6d6af3491f2835a4aaf2b5b7649e1983d7b17c93b4302228e7653d +size 227009 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YW-0MMPGAD7k.flac b/Foundation/Audio_Grounding_AudioGrounding/YW-0MMPGAD7k.flac new file mode 100644 index 0000000000000000000000000000000000000000..cd1f59f611aea2ea5e21c81ce7d86726a7f4c796 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YW-0MMPGAD7k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d558860772c0aebcd29efa2898fa0bb2ac11a7dea1242ec57a9236f7f8386753 +size 249909 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YW7OJevEgq7w.flac b/Foundation/Audio_Grounding_AudioGrounding/YW7OJevEgq7w.flac new file mode 100644 index 0000000000000000000000000000000000000000..ebae41ada9f42d3fcc113aed7a60627d80dee3e4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YW7OJevEgq7w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eceb065bca629f78c3d50564b95d01e2a6175643bc29ac036206afbe8700f06c +size 193984 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YWMwYUbpKzfY.flac b/Foundation/Audio_Grounding_AudioGrounding/YWMwYUbpKzfY.flac new file mode 100644 index 0000000000000000000000000000000000000000..1b91571e3ae073a2be5064ff7f6051bcfe46d26e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YWMwYUbpKzfY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3474b559597e28b5c931180e0aef41667f4d235d14ab5171349def19f44e0e3f +size 215428 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXG5f9OyzO_A.flac b/Foundation/Audio_Grounding_AudioGrounding/YXG5f9OyzO_A.flac new file mode 100644 index 0000000000000000000000000000000000000000..5690c0eeff999e7baf03e220057e2ed87a1f0e1c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXG5f9OyzO_A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5cc3df4e1aa067f39be1c9793e3c009403ada61eed928c11b35165d40ef48ce8 +size 259257 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXL8JV9qXGLE.flac b/Foundation/Audio_Grounding_AudioGrounding/YXL8JV9qXGLE.flac new file mode 100644 index 0000000000000000000000000000000000000000..8562aed00285c487e8e73ca12ca3c7dae8354de6 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXL8JV9qXGLE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ceb64c1980cfaa5570046ee071a0d76acdbdd808346cc7aa08347299be00ac72 +size 253518 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXTGcypOsLgw.flac b/Foundation/Audio_Grounding_AudioGrounding/YXTGcypOsLgw.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5e60bbb0f111692b616e46caf23c91f0b22a611 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXTGcypOsLgw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:60f6a2c1fcfc5e7c08074804864b50f726abdb0a00a3221def7174da49ce3918 +size 267168 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXVlnx4ofu2k.flac b/Foundation/Audio_Grounding_AudioGrounding/YXVlnx4ofu2k.flac new file mode 100644 index 0000000000000000000000000000000000000000..43360c0190dccb09a603a1fdbe9e0c41a3a6c22e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXVlnx4ofu2k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a08b1eadda8804fd58924396bfff7e76c1bfefba0afbd848181d7f8ae8fa4dc +size 254568 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXi6V0LGvqoo.flac b/Foundation/Audio_Grounding_AudioGrounding/YXi6V0LGvqoo.flac new file mode 100644 index 0000000000000000000000000000000000000000..5540daa93f9b5ef2308ce5b00479a602e5b862e5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXi6V0LGvqoo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6e99cd8d9c506ee7c84963d36358909a0ecfe388ef40f6ba8f5c213e21347fe +size 258149 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXm0LegZNUOQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YXm0LegZNUOQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..080e96c3a2aa012b415bd58404871462f31de428 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXm0LegZNUOQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8c26443f60cae15b1c3ffdea01a39d5897083af3736c82f462c0c9f55de812f8 +size 167334 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YXn4SruTQFfY.flac b/Foundation/Audio_Grounding_AudioGrounding/YXn4SruTQFfY.flac new file mode 100644 index 0000000000000000000000000000000000000000..c06544729e715474948e47c6b3770e1103ae7e31 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YXn4SruTQFfY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1652e4efb9688b67c0bb86c33e5a35447b14ebd8e230005eedc6bcdb68b07a14 +size 260237 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YYZOEAWhD7js.flac b/Foundation/Audio_Grounding_AudioGrounding/YYZOEAWhD7js.flac new file mode 100644 index 0000000000000000000000000000000000000000..e419df746758592b237c7335ce51e546988d3005 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YYZOEAWhD7js.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9ed1a2d3cb044fde7e4eee8120fc8a09f0cbac5beca26942fce9a3c80e03cb2 +size 263972 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YYp6CI7K4UcY.flac b/Foundation/Audio_Grounding_AudioGrounding/YYp6CI7K4UcY.flac new file mode 100644 index 0000000000000000000000000000000000000000..4dd0c1771a32c48f6c77c7bea77d0c0619b600a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YYp6CI7K4UcY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:331c486bf168b858b694cab2aae9d4a46d7b9cc5321fd1cbc6546aa716d2bb51 +size 205045 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YZG4byeZq9S4.flac b/Foundation/Audio_Grounding_AudioGrounding/YZG4byeZq9S4.flac new file mode 100644 index 0000000000000000000000000000000000000000..d69b2f62f569e240482275b4482a4fd5f09d11d5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YZG4byeZq9S4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c001c60408a8ac2cd36f7959d749ec169a73c44d7ee675596f09351b97d73a04 +size 225318 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YZK1kWFoHLQo.flac b/Foundation/Audio_Grounding_AudioGrounding/YZK1kWFoHLQo.flac new file mode 100644 index 0000000000000000000000000000000000000000..192286e588685ebd7f806888aa138c16369daceb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YZK1kWFoHLQo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:763cb88f213c2bdebb6b7386d192c97f50d236b62b155a718a654f3929b627d4 +size 197877 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YZLHtRWod3wM.flac b/Foundation/Audio_Grounding_AudioGrounding/YZLHtRWod3wM.flac new file mode 100644 index 0000000000000000000000000000000000000000..5aa7010da4dba03878e1adb0a4603e04c0873e1b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YZLHtRWod3wM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:45cbbc752ef6b366593963b8a02b160f93169cff7ce7788e34e4543869045d69 +size 278656 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YZVlZro7vrxw.flac b/Foundation/Audio_Grounding_AudioGrounding/YZVlZro7vrxw.flac new file mode 100644 index 0000000000000000000000000000000000000000..9d777c89871145019d14eb2e786a0a182f294e75 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YZVlZro7vrxw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:514052e9c60343e713f20cf6463f0d21a16dd831d7aee9821184e34ebddd7b09 +size 231456 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YZq8UkUAhLZQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YZq8UkUAhLZQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..33f620021acd2fd1b674dd0f05f49ba0c34c091b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YZq8UkUAhLZQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7fec1bba88796ce1f4e7a93640f283ab5b2d344170c0cf0b7e4d5bbc2e6cbca5 +size 283273 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y_5DSfZnQZPw.flac b/Foundation/Audio_Grounding_AudioGrounding/Y_5DSfZnQZPw.flac new file mode 100644 index 0000000000000000000000000000000000000000..0de23e00b5342b06a2a879aaa9316880915d422a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y_5DSfZnQZPw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae3bed872e8863b0fedf6a59ef88b2d7a9009dc4fc52bc92d3adc531206cb7eb +size 219757 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y_GI7meqlYZk.flac b/Foundation/Audio_Grounding_AudioGrounding/Y_GI7meqlYZk.flac new file mode 100644 index 0000000000000000000000000000000000000000..6c636c85160f42c464986d69014c63128ee4f142 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y_GI7meqlYZk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5633112f221449218668afa0e8dfdd4e0aa75474b5467ee2b9eea17bbf6fa52a +size 226444 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y_SHrBNEes38.flac b/Foundation/Audio_Grounding_AudioGrounding/Y_SHrBNEes38.flac new file mode 100644 index 0000000000000000000000000000000000000000..28ef16edd3f483508c1a92f7050265696660d233 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y_SHrBNEes38.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:234c4c479998156da9821dbc85bb4fc0192fade241b5832c7d71311a20a84c44 +size 194213 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y__a3v2k4Svg.flac b/Foundation/Audio_Grounding_AudioGrounding/Y__a3v2k4Svg.flac new file mode 100644 index 0000000000000000000000000000000000000000..4b21bf45fc59ba8c268a9b97f61bfc1bacbcc079 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y__a3v2k4Svg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c51febd113ed9ceb6edf409f0b09f4ab62881556b79e1858ef2a58250a84595d +size 153841 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Y_kL6LQF41cA.flac b/Foundation/Audio_Grounding_AudioGrounding/Y_kL6LQF41cA.flac new file mode 100644 index 0000000000000000000000000000000000000000..28e7f70322399ff72e2213c2f07602d4b7c1d1c5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Y_kL6LQF41cA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6152830f0fbcdebe8431be6aecd5d203c3181eebdcf3970e6f6f64c89277f196 +size 138490 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YaByChd13yEA.flac b/Foundation/Audio_Grounding_AudioGrounding/YaByChd13yEA.flac new file mode 100644 index 0000000000000000000000000000000000000000..c3184cf7dbb746e7d98076ec9521ac044636ec0c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YaByChd13yEA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d7af10066f245222da08462d0933c2c7fecb8ea234b86ca26995e080c213e4d +size 237698 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YaC_Ynuy_cl0.flac b/Foundation/Audio_Grounding_AudioGrounding/YaC_Ynuy_cl0.flac new file mode 100644 index 0000000000000000000000000000000000000000..cbc7ba02a37d979d1e7947d4df83b98c1616b37e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YaC_Ynuy_cl0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c5f3ece24383a944fc601c70b37595f476a713447a5bd59be20e85ad0156550 +size 231854 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YaWhMvg9JbxM.flac b/Foundation/Audio_Grounding_AudioGrounding/YaWhMvg9JbxM.flac new file mode 100644 index 0000000000000000000000000000000000000000..1024a1369080e2d858805eeec05ef4fa2152ea3f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YaWhMvg9JbxM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a752c633e52c424d13976cadb038f3fb1b5b334e671fd3a8fd7340e93ee46da +size 230113 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YaoK7MLknLCk.flac b/Foundation/Audio_Grounding_AudioGrounding/YaoK7MLknLCk.flac new file mode 100644 index 0000000000000000000000000000000000000000..7ee3c1a811ac28c8004495a89e2c967be28d5775 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YaoK7MLknLCk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fd7720fe53ce3e4cbf6449fdcabcc069b66b2968e38753b3cc399ab1f90ac30 +size 209818 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YarFwKX6iU38.flac b/Foundation/Audio_Grounding_AudioGrounding/YarFwKX6iU38.flac new file mode 100644 index 0000000000000000000000000000000000000000..8ed7dbecb679409f1a3402553fcf9222ceae243e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YarFwKX6iU38.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a541b93225794ac2bf4372ca2cb8ce6f851f583348cf3f71023ba936d362e460 +size 252377 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YayUqgnURxgo.flac b/Foundation/Audio_Grounding_AudioGrounding/YayUqgnURxgo.flac new file mode 100644 index 0000000000000000000000000000000000000000..12e8dd1213008c72ac0c0bee05d4d9f1a0021c0a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YayUqgnURxgo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:256c87b8b271e0f6b81caae5a6ec8b91dff7ca8f2975a2d4e7827a474042f964 +size 241229 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yb4xHgy3Fv3g.flac b/Foundation/Audio_Grounding_AudioGrounding/Yb4xHgy3Fv3g.flac new file mode 100644 index 0000000000000000000000000000000000000000..f6b6af4ca42e7555e9e60f5c58fef7c48f4ec67f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yb4xHgy3Fv3g.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77cc29306c5dcda5e5b7e6ad5bda098d1cb353800731f2b89c4002a1e70391c2 +size 240654 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YbPOo7UBbwOU.flac b/Foundation/Audio_Grounding_AudioGrounding/YbPOo7UBbwOU.flac new file mode 100644 index 0000000000000000000000000000000000000000..d326677b7371a1f1b097754b426742811fbce733 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YbPOo7UBbwOU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3aac6aaeacb9487d9e472774c2806e4443fb3740746bf45d5d1d89340164917d +size 234177 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YbS4flke6PUg.flac b/Foundation/Audio_Grounding_AudioGrounding/YbS4flke6PUg.flac new file mode 100644 index 0000000000000000000000000000000000000000..55f47bf2b6058810a55122b6b7c9fc1442a02e8c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YbS4flke6PUg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6884c062bacef5293c9eacc93c7db3ed75aa9e189d1c2e46ff3e8267200f7c8f +size 253666 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YbWYqngojj7Y.flac b/Foundation/Audio_Grounding_AudioGrounding/YbWYqngojj7Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..5ea2c803cc3752aa3850daeb34a5242cbbe2b5a8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YbWYqngojj7Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bccc30d4c21b88ea1d2af4453dc08d0be04852e0651ea419a0fb96a81914854e +size 205914 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ybpv_LneHmfU.flac b/Foundation/Audio_Grounding_AudioGrounding/Ybpv_LneHmfU.flac new file mode 100644 index 0000000000000000000000000000000000000000..c8ad7c2b3dd3ece55de4b9b6002e1878fd989771 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ybpv_LneHmfU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67cc8bad9b122e00a64040ad066af31877634c5e4a529c1d93b79bd754d9f5e8 +size 288263 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YbygBWUkpaC8.flac b/Foundation/Audio_Grounding_AudioGrounding/YbygBWUkpaC8.flac new file mode 100644 index 0000000000000000000000000000000000000000..406d03fb3e286c1c41e73fd967b20a9d854702cc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YbygBWUkpaC8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df663f0afc5ce5acc949a05c16419a1fd5c3da8ab4089a2f5e99cea69aa62a0f +size 213084 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yc3nlaAkv9bA.flac b/Foundation/Audio_Grounding_AudioGrounding/Yc3nlaAkv9bA.flac new file mode 100644 index 0000000000000000000000000000000000000000..673bd092a3a8576fb82af8c80769388228580e71 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yc3nlaAkv9bA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c07a1bff6353db83940d03d8cff949da44d930c1c512b9a87fd7776be7848076 +size 181894 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcE7sLRthlZM.flac b/Foundation/Audio_Grounding_AudioGrounding/YcE7sLRthlZM.flac new file mode 100644 index 0000000000000000000000000000000000000000..1e8d11bc63409456d96d1066bfdbf02e7aa84be8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcE7sLRthlZM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6827a84f57faea6c4009cd31ed51ae47e07b9bed9b2062ffd1dcbc342492d1dc +size 247882 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcHm9Hdis3wE.flac b/Foundation/Audio_Grounding_AudioGrounding/YcHm9Hdis3wE.flac new file mode 100644 index 0000000000000000000000000000000000000000..5f24d4f7f508a1d1d54ec613a10d7c5325641682 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcHm9Hdis3wE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37ba304a8043799c49b19c62d50f2e0fe2058363dbc5f17b489e8508206cd6f4 +size 211095 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcLtItZ0WGQc.flac b/Foundation/Audio_Grounding_AudioGrounding/YcLtItZ0WGQc.flac new file mode 100644 index 0000000000000000000000000000000000000000..babb93508ca6a4e8525e504c3cb2259610524950 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcLtItZ0WGQc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd6c3792790b05943618b9ee5528c7523adf7f9a4d4767f9e8aeb4834d4d4d27 +size 243446 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcNn1rxrQq9c.flac b/Foundation/Audio_Grounding_AudioGrounding/YcNn1rxrQq9c.flac new file mode 100644 index 0000000000000000000000000000000000000000..9cb18fd4c6b55860702c3e0eb3a53510f6c770e9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcNn1rxrQq9c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c14f311567f5b63a2356c291048c435704a7d8a09f2386117b7666d53c0a61b3 +size 272610 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcO4_J-DNZvM.flac b/Foundation/Audio_Grounding_AudioGrounding/YcO4_J-DNZvM.flac new file mode 100644 index 0000000000000000000000000000000000000000..f0eedcb1ed01f9107c845a710c546ccdb66daae0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcO4_J-DNZvM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cab334c60eac5e1b55370ef251f5827faa13ed25336f18a893c1d13a03370b7e +size 269073 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcSURZuti4dU.flac b/Foundation/Audio_Grounding_AudioGrounding/YcSURZuti4dU.flac new file mode 100644 index 0000000000000000000000000000000000000000..8c80907bb8763e3b233b27319c6a108e64e14f7b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcSURZuti4dU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f44a59b48fdfdc8b93e2eff265346c1ce8261f9e335ea17076c1ab03d8f1a85a +size 189869 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YcnBAAZDhuVA.flac b/Foundation/Audio_Grounding_AudioGrounding/YcnBAAZDhuVA.flac new file mode 100644 index 0000000000000000000000000000000000000000..69e0d1b79983031282d9cb30e2d7503840b7d6f7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YcnBAAZDhuVA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dca65811fb61a524a81b973638847d929db23034ef6a20b3bc233300526d8a86 +size 268648 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yd55UQEWwDN8.flac b/Foundation/Audio_Grounding_AudioGrounding/Yd55UQEWwDN8.flac new file mode 100644 index 0000000000000000000000000000000000000000..4d04925d2d6cef618124a08fa3f30c90ad7ad1b3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yd55UQEWwDN8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01412145edc4666bde653ab6421bc1fd994626c0c54e4a250487931a5ce13f65 +size 290873 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YdAepTxzUYFQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YdAepTxzUYFQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..fb6043561ed95a027fd8fe0bdef202494097d2f9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YdAepTxzUYFQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1ebf4c3c8c852baaaf99552b74344c532751d805abf551e993cf1a96315876a +size 214727 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YdHce3m03dgk.flac b/Foundation/Audio_Grounding_AudioGrounding/YdHce3m03dgk.flac new file mode 100644 index 0000000000000000000000000000000000000000..b265e6eb13badb713ace604fbe940d594dc0d48a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YdHce3m03dgk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a8ec796fbb121db783a82da998375c5e70004e0f52c2109780923b297b08819 +size 165352 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YdKCJpCevfCI.flac b/Foundation/Audio_Grounding_AudioGrounding/YdKCJpCevfCI.flac new file mode 100644 index 0000000000000000000000000000000000000000..05fd194334d95e6fc21fd58bf7c7cb99ee27c15d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YdKCJpCevfCI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e052b338722ed87c545c41b8697fd7a209bc8dd90ebbc851be06f081ff376aad +size 239250 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YdgnWKrBl4OQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YdgnWKrBl4OQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..d247eafd728a4ed0c48319701e29c234cfc83659 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YdgnWKrBl4OQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:904db99adeb99efbe504c662b12f6bd772d0c7a620a2a86444b5aa81e9768751 +size 174115 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ye30uwJvj4Io.flac b/Foundation/Audio_Grounding_AudioGrounding/Ye30uwJvj4Io.flac new file mode 100644 index 0000000000000000000000000000000000000000..7472ebf19855ef7801a3f3299d52970098b5259a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ye30uwJvj4Io.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d845c5efecd9af9ac3a0a6d206a3f2d80ed98cce069d56c40546cc4c37b89b13 +size 117903 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ye6y_ejwxrng.flac b/Foundation/Audio_Grounding_AudioGrounding/Ye6y_ejwxrng.flac new file mode 100644 index 0000000000000000000000000000000000000000..e609cebc13eec6658bf9b0fdbc08c743391c1973 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ye6y_ejwxrng.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3248ca8835b09b423435fb809ba418165a2b96283938f0aea5bbce84cb7829d7 +size 219803 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YeAroMk9QxuY.flac b/Foundation/Audio_Grounding_AudioGrounding/YeAroMk9QxuY.flac new file mode 100644 index 0000000000000000000000000000000000000000..0924bbdc0b37d949513a2ebeee6901280923cd57 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YeAroMk9QxuY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e44247e8c0431f109c0e7055a7ec279c6089929965252fed7544f5ce8fdfbfc +size 266308 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YeTok2QICK6Y.flac b/Foundation/Audio_Grounding_AudioGrounding/YeTok2QICK6Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..5a5bc36ab5329230af9a8db77fdbcd12af7b89a8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YeTok2QICK6Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:80f9ea0e67edf81f36ff0b73025e8983e3442bb5d2816e85b23cd5e14f47802c +size 240428 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YelztUCeNQvQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YelztUCeNQvQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..603056bf65b2c9e1bec289a570d67bb98a5ac26d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YelztUCeNQvQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3ce099628400de55e28fdf58a737ad93fb07bfb0e4320df6658b373c47820e02 +size 250072 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yet5VaXFQKzs.flac b/Foundation/Audio_Grounding_AudioGrounding/Yet5VaXFQKzs.flac new file mode 100644 index 0000000000000000000000000000000000000000..a9e6c1b8e310f888e1bb196f64d834130a37692a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yet5VaXFQKzs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5bc8c9099a1836abbf6f1c6f31254fdb6fa29ffff706eb176a4b61f988abd027 +size 208027 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yf98tFMTAOrQ.flac b/Foundation/Audio_Grounding_AudioGrounding/Yf98tFMTAOrQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..21003f483b49bb089e3f223bef2d5f3fe44cfada --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yf98tFMTAOrQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:07a5b456a96202c94f796d9d9ee2a2db3664666bd199c2e92f32a2ca8d1581c9 +size 216792 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YfKwl4gOlPE4.flac b/Foundation/Audio_Grounding_AudioGrounding/YfKwl4gOlPE4.flac new file mode 100644 index 0000000000000000000000000000000000000000..0fdbe50413a064827f64350714ad4c455425ab4b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YfKwl4gOlPE4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:24b325d0b163251061d26e3976ca0cdb5ac781a5380b02dc0c38c1ad181a2c95 +size 205848 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YfM10Whc-ro0.flac b/Foundation/Audio_Grounding_AudioGrounding/YfM10Whc-ro0.flac new file mode 100644 index 0000000000000000000000000000000000000000..176cdc4d79b6a1902c5fe515d81f1cf0e7bbb0e2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YfM10Whc-ro0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0480d99956b694d4d0408d5ad915a70ecf8efe07f552d56835e8459d7397aeb5 +size 298564 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yfg_unNt1yds.flac b/Foundation/Audio_Grounding_AudioGrounding/Yfg_unNt1yds.flac new file mode 100644 index 0000000000000000000000000000000000000000..026b7083ec85d4856df9fc1766bef133bd5fff05 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yfg_unNt1yds.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c7d1c9dab639bc940e24f79c2ad5af053cc433ebece98e2487cd401d7543c10 +size 250593 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yg26ODac1y3o.flac b/Foundation/Audio_Grounding_AudioGrounding/Yg26ODac1y3o.flac new file mode 100644 index 0000000000000000000000000000000000000000..789d9ddd924d6f1a3eec02bc2b007ec291d944eb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yg26ODac1y3o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc657abe2163d3b62b8af17919c54bdf0a56998acaefd375729430574c2321f9 +size 193922 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgA6n2-9zxKM.flac b/Foundation/Audio_Grounding_AudioGrounding/YgA6n2-9zxKM.flac new file mode 100644 index 0000000000000000000000000000000000000000..244436a42f313954d80ca0dc47048364591f9a2f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgA6n2-9zxKM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0f91db11ce9514654740e3848a1d05024dd2cb2bba46579715342980df176ea +size 264669 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgIdmmdHdAJ8.flac b/Foundation/Audio_Grounding_AudioGrounding/YgIdmmdHdAJ8.flac new file mode 100644 index 0000000000000000000000000000000000000000..aea97a9fdcc60a7e22e8714077159e1ed0a38cfa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgIdmmdHdAJ8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:58ae950d7b85e04e64d19a613716940f0198e8f01694f90565eb007b1ca8bfe8 +size 188418 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgJ_qu2EC8Kw.flac b/Foundation/Audio_Grounding_AudioGrounding/YgJ_qu2EC8Kw.flac new file mode 100644 index 0000000000000000000000000000000000000000..4ce3ec96efd921820f2e723e463a3d6d597f4520 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgJ_qu2EC8Kw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c4f457d1a44de1ab3e602229e406482bca01192d5704f1f019ea61d111a98ec +size 268204 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgWVY03cqNTA.flac b/Foundation/Audio_Grounding_AudioGrounding/YgWVY03cqNTA.flac new file mode 100644 index 0000000000000000000000000000000000000000..b16e68041a29b58be0aa187d64f25a557ee60baf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgWVY03cqNTA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09f4c627be504ec446e01eef02fc5493a97144664361c0ef9bd621b12f72c5d5 +size 248213 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgXJXfZra9TA.flac b/Foundation/Audio_Grounding_AudioGrounding/YgXJXfZra9TA.flac new file mode 100644 index 0000000000000000000000000000000000000000..c942e4890f66de74e1fe5674d82ff3badbe4d13e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgXJXfZra9TA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3dc6e79d588d45f07994d2cab4231b00e8749420b243ac40d475935cae427b7 +size 323812 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgXU3IbRXYIQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YgXU3IbRXYIQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..7f57f3091cea7bb1ea6a158b85345e8cac33c52c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgXU3IbRXYIQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfb85feab7ebc5b106450c792ed11b5454d68f6ab61d64ce5b2b2dddb4423028 +size 187025 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgYDoraoD85E.flac b/Foundation/Audio_Grounding_AudioGrounding/YgYDoraoD85E.flac new file mode 100644 index 0000000000000000000000000000000000000000..52c41e8f937c66413df37c8a6b515728f1230dea --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgYDoraoD85E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81014a2a1126120c4c1dfa9255c536860729fa8ea776e9d71bffd58fd018a23a +size 251027 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgZAnkBJMK5k.flac b/Foundation/Audio_Grounding_AudioGrounding/YgZAnkBJMK5k.flac new file mode 100644 index 0000000000000000000000000000000000000000..86b27fcdce03efc45c2d40010e36409a98089982 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgZAnkBJMK5k.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ac46cfb7899470f010be7407c1661bf282fee4defdde829a17c3cd6e4a0b37c1 +size 142529 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YglyZUCN3BfU.flac b/Foundation/Audio_Grounding_AudioGrounding/YglyZUCN3BfU.flac new file mode 100644 index 0000000000000000000000000000000000000000..1443484198acd5c221726536bda2e930897840e0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YglyZUCN3BfU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f497b11733dc9cb2ee08bae12453e730217be82b9da6686044969268ae8b1999 +size 273365 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YgqYEFSj6vBA.flac b/Foundation/Audio_Grounding_AudioGrounding/YgqYEFSj6vBA.flac new file mode 100644 index 0000000000000000000000000000000000000000..1a2db0a5f6beb907f333318148764bb4819bee57 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YgqYEFSj6vBA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6569f69c04eb1ccc0a39b5961a21070ff105045b1276d6615256d8510c0a268a +size 269100 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YhFHkJ2qWm90.flac b/Foundation/Audio_Grounding_AudioGrounding/YhFHkJ2qWm90.flac new file mode 100644 index 0000000000000000000000000000000000000000..c4f35d3a0a810f8ca22a5f67671b5e6c6e4671be --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YhFHkJ2qWm90.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:499760039c07bd139d2d3ba7ca5be119e1e6ddd70ab7e06a2940f3872115bbc4 +size 250146 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YhogpqdigubM.flac b/Foundation/Audio_Grounding_AudioGrounding/YhogpqdigubM.flac new file mode 100644 index 0000000000000000000000000000000000000000..c3ff28097e1d1133ecfd5f51eea4417f23c3d79f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YhogpqdigubM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ebe2460e737e19c40d25d1380e2f3fa347ac1d59d22b29665be77c1edcc89063 +size 284010 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yi2tHuOnJTWY.flac b/Foundation/Audio_Grounding_AudioGrounding/Yi2tHuOnJTWY.flac new file mode 100644 index 0000000000000000000000000000000000000000..d5219cbb38fa2966028bedc9f296228ac6ac8b67 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yi2tHuOnJTWY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3c38d7b0f26f27c4f2818186cb23c98516ce11564b26569e78db449702bcd6c3 +size 279369 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yi9RHYhT0y0I.flac b/Foundation/Audio_Grounding_AudioGrounding/Yi9RHYhT0y0I.flac new file mode 100644 index 0000000000000000000000000000000000000000..69098a0159560dce64e2ba360773df497a5c6eb2 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yi9RHYhT0y0I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2e2e6c19bdf2706a120723fe0128134cb9ad09b760941474b3681021e620ea4 +size 130840 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YiIASq403bOQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YiIASq403bOQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..91824c1967b7b2ced6328239bcb3dbf8245a2844 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YiIASq403bOQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39391bc68b4f1c9bead2e2bbf6721711069280568f54dc944fdfcf064a3b7829 +size 212461 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YiQGztcbPgWI.flac b/Foundation/Audio_Grounding_AudioGrounding/YiQGztcbPgWI.flac new file mode 100644 index 0000000000000000000000000000000000000000..fb35229e165b045796df5dbc416416e5fa464628 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YiQGztcbPgWI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9cef05bb996445976086ade4a5ffc17462c2d4077ec99ef49c57afe04f361c6a +size 279543 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YikhBs__x9xk.flac b/Foundation/Audio_Grounding_AudioGrounding/YikhBs__x9xk.flac new file mode 100644 index 0000000000000000000000000000000000000000..0f59b38d618a0583e3cb60281684b4ad1d69df18 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YikhBs__x9xk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a07970bf87286c76b40a8520002514053aac20c24311a886e72cbb63d112eed +size 227864 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yimbp68i3pxo.flac b/Foundation/Audio_Grounding_AudioGrounding/Yimbp68i3pxo.flac new file mode 100644 index 0000000000000000000000000000000000000000..eb2b340df7a0fc790f5416f366ce6a1d6fd0dbf0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yimbp68i3pxo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9eb755a31e6f57dcef930173210c42ee864b0103e60e852810b0d31bf70e307c +size 245067 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YjCExyUac_xM.flac b/Foundation/Audio_Grounding_AudioGrounding/YjCExyUac_xM.flac new file mode 100644 index 0000000000000000000000000000000000000000..25ac078554399afab0f42dbb2a69c54cfffd6ed5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YjCExyUac_xM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:735287aacec5f4a9f0e707d26c82ebe6ca589689d6fcdb7ce734f13ad9bc6d10 +size 282831 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YjHv726cD6TQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YjHv726cD6TQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..948f582d160447457c5f5a7dd9d389438d128ee4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YjHv726cD6TQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:963e162369d239160bf8d7370b4c155bb17e83a68cdae0aba48ba8bcb15dfaf3 +size 232057 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YjOgB2_9y9S8.flac b/Foundation/Audio_Grounding_AudioGrounding/YjOgB2_9y9S8.flac new file mode 100644 index 0000000000000000000000000000000000000000..c67c6ef93b1f7efcac6778f34eb705051b8e7f59 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YjOgB2_9y9S8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3a95abf89664e759d914474c32d25b3464f55b63b156e662cfd91836239f11b +size 182620 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YjbIEjeGUpMc.flac b/Foundation/Audio_Grounding_AudioGrounding/YjbIEjeGUpMc.flac new file mode 100644 index 0000000000000000000000000000000000000000..8709fd9aa018885ca22be5def1b0419ee3b408f4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YjbIEjeGUpMc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b8ee66806f1bb392172af163263604eae1334418cefea74dd3b60ee5e9567f4 +size 64326 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YkC2t97-pw5Q.flac b/Foundation/Audio_Grounding_AudioGrounding/YkC2t97-pw5Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..7d0c1ed3a8d42eeba3eeb30598865a1c7ff69569 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YkC2t97-pw5Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a85a0d42eb8476f124617389e90a9461d8a3eb1da3baba6bafd32c46f1b9189 +size 254686 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YkFyDYvCvOqo.flac b/Foundation/Audio_Grounding_AudioGrounding/YkFyDYvCvOqo.flac new file mode 100644 index 0000000000000000000000000000000000000000..df977997235a338bb878dcda537338fd43752c01 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YkFyDYvCvOqo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:230d8ef7f45705797950f1b9c159fb34f0915066cd8b39b3a74c17c1b78db42a +size 225812 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YkPfjp_C66HA.flac b/Foundation/Audio_Grounding_AudioGrounding/YkPfjp_C66HA.flac new file mode 100644 index 0000000000000000000000000000000000000000..6502a4044be92bef33c8e468689109c8900b6673 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YkPfjp_C66HA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8d1b49f7e9e0915d18e484805019ef98e18897b63f54c036e4adfe0c4d17f3b +size 243498 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YkY1wpO7_pMs.flac b/Foundation/Audio_Grounding_AudioGrounding/YkY1wpO7_pMs.flac new file mode 100644 index 0000000000000000000000000000000000000000..3b3514fb93f2c595e163c5d6780c18b15bb96797 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YkY1wpO7_pMs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae7e1a698bc6fbfc53fbe3829ed2844da75de62ab96c405f6310ce60b5919d64 +size 188854 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ykk_Q4naP_9o.flac b/Foundation/Audio_Grounding_AudioGrounding/Ykk_Q4naP_9o.flac new file mode 100644 index 0000000000000000000000000000000000000000..856978e8f29e0069ddb8602c0dd3c0d8d9c0c7cb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ykk_Q4naP_9o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a0a71f396d8d804620adcf1bda5233dad75d8bb2c647d8defc86c3f873af204b +size 264975 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yl96ZcCdYwfY.flac b/Foundation/Audio_Grounding_AudioGrounding/Yl96ZcCdYwfY.flac new file mode 100644 index 0000000000000000000000000000000000000000..b403c18731c2ca6b86e8d59950a06b1fadcb291b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yl96ZcCdYwfY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23776c0d9d818f2a75920fae5df2d97c988fc261afe9181e86216138cf8d5945 +size 269073 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlDtptH9Odf4.flac b/Foundation/Audio_Grounding_AudioGrounding/YlDtptH9Odf4.flac new file mode 100644 index 0000000000000000000000000000000000000000..83bdc753b2ea50b5bdeb55a527510cc13a419654 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlDtptH9Odf4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:444b31c16e99e73964a8e4a9a85175f0e96f63233991c45a18604e4ce7b4e661 +size 176736 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlJ6fEVn7IlM.flac b/Foundation/Audio_Grounding_AudioGrounding/YlJ6fEVn7IlM.flac new file mode 100644 index 0000000000000000000000000000000000000000..58232d253f0f84acab398af476c8c01fd83136c4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlJ6fEVn7IlM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6681836d109c7fb7f102ddb98830031f174d6fbeec57ddc6f4fa93e5c60515ec +size 176788 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlTyPKec30Bk.flac b/Foundation/Audio_Grounding_AudioGrounding/YlTyPKec30Bk.flac new file mode 100644 index 0000000000000000000000000000000000000000..d865b07415c817a790047c2b5c69f2917e595c0f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlTyPKec30Bk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:10c91bd18cb40240c090f8e9695f18f1dd492f220998bf9d4d221fbe62669474 +size 245976 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlX3k5p2I_g0.flac b/Foundation/Audio_Grounding_AudioGrounding/YlX3k5p2I_g0.flac new file mode 100644 index 0000000000000000000000000000000000000000..24c7abc29c1a14b45230b8f7d966607967e7b12f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlX3k5p2I_g0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dc568c2197196eef75d5b2c0855791cb155d25c86a485e2535f749f252f8d134 +size 244856 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlZ7B2De3kqs.flac b/Foundation/Audio_Grounding_AudioGrounding/YlZ7B2De3kqs.flac new file mode 100644 index 0000000000000000000000000000000000000000..b5225d3ed71edcfbfc08fd8f7500cc646d1e843d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlZ7B2De3kqs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bdb269a4f0b0e2bbf0915853f03eaadd50bfa762459ef527e6ea7b38236192dd +size 241147 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YlemaMeWsO1s.flac b/Foundation/Audio_Grounding_AudioGrounding/YlemaMeWsO1s.flac new file mode 100644 index 0000000000000000000000000000000000000000..3e522a0323687e54e6801da63cb8085cb75cc0f3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YlemaMeWsO1s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:36ca092ef2b075bc51829fa5cb0120b7813b2a2d434d891a4ce4285f9257d0b1 +size 245777 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ylywxhiw977o.flac b/Foundation/Audio_Grounding_AudioGrounding/Ylywxhiw977o.flac new file mode 100644 index 0000000000000000000000000000000000000000..f4413d4d754f95bd9a3f617b5637db2bb4b146cf --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ylywxhiw977o.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b776c631825e4f2b123b59aa8b6203b6208a7b12a92a53743caa70fc07396e7f +size 280984 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YmHT7IM34WTY.flac b/Foundation/Audio_Grounding_AudioGrounding/YmHT7IM34WTY.flac new file mode 100644 index 0000000000000000000000000000000000000000..90a04c5af46ba48129fc9962a33fff658ddc75f5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YmHT7IM34WTY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2180fd19c99e31b0594b97fcef37d30fb07c89b8ff75b20731d43b2b8e3e2a85 +size 217641 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YmK-KS9OfG0s.flac b/Foundation/Audio_Grounding_AudioGrounding/YmK-KS9OfG0s.flac new file mode 100644 index 0000000000000000000000000000000000000000..519bb6e7dd22465289f75a759354e8216b078b79 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YmK-KS9OfG0s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64a4d76dbe0c03e9b49ae43e68ad2b4535dc9a42ea0363408f5fcf1c436ed361 +size 195875 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YmKaqk8Fc4Kc.flac b/Foundation/Audio_Grounding_AudioGrounding/YmKaqk8Fc4Kc.flac new file mode 100644 index 0000000000000000000000000000000000000000..491e64cca02b3f7b9f04549fac23adc4cf02ee2a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YmKaqk8Fc4Kc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d2dfd566b696c924c0a3e1d0217712910c1504e0c22187c886bcc73283b6403 +size 235757 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YmUpVz-n8FqQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YmUpVz-n8FqQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..29e45dc181404aa1e3706e17cc096657056914f5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YmUpVz-n8FqQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7ff066699f02232c55028cd6b72f70e5abb3bbc2e951e062395a39a02a38ab3 +size 261163 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YmvbjcYC2TaY.flac b/Foundation/Audio_Grounding_AudioGrounding/YmvbjcYC2TaY.flac new file mode 100644 index 0000000000000000000000000000000000000000..04f7c414e7854de340e092cf8b624d8ac1481558 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YmvbjcYC2TaY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:30a704a79f03707c9d22df8467620f949289339f24ad646fbef8e18fb7e40704 +size 219823 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yn7CbU6TM7Lw.flac b/Foundation/Audio_Grounding_AudioGrounding/Yn7CbU6TM7Lw.flac new file mode 100644 index 0000000000000000000000000000000000000000..bacfe372006ffa5c586abb84fb727b46d0ff115c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yn7CbU6TM7Lw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7de73fe4161308e324b955093d7b17fceaa72b0e2d7c1c80f9c0ca20d93a6abf +size 209601 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yn7IFnrK8POM.flac b/Foundation/Audio_Grounding_AudioGrounding/Yn7IFnrK8POM.flac new file mode 100644 index 0000000000000000000000000000000000000000..c99038031795efb134112e3da447c87026132312 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yn7IFnrK8POM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c512cd2928caf29d3b6881d8238b9677fe1aabf53dc160893eecf5005db776b +size 288267 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YnKQ6BTfiAR8.flac b/Foundation/Audio_Grounding_AudioGrounding/YnKQ6BTfiAR8.flac new file mode 100644 index 0000000000000000000000000000000000000000..97f9dcbc94f93e1012d76e16c482410426459936 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YnKQ6BTfiAR8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1fcf7aaf92461a5a5b99ed4ab1442e82b2b3eeb1cd2a30f0ff2019e0d5a295c3 +size 247402 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YnQy5YuL-W6A.flac b/Foundation/Audio_Grounding_AudioGrounding/YnQy5YuL-W6A.flac new file mode 100644 index 0000000000000000000000000000000000000000..1ef64377ee8f6824432609e9f61fc2886a78ce10 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YnQy5YuL-W6A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:84d4f7923b3f07a8f8f6cb798b92c915c630ae07bd2cad537d56227ff419a0bf +size 237269 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YnWEbiJc6zxc.flac b/Foundation/Audio_Grounding_AudioGrounding/YnWEbiJc6zxc.flac new file mode 100644 index 0000000000000000000000000000000000000000..02e0fd328f4da01ca9c697d486ea74b164acf1b1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YnWEbiJc6zxc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:767f9454619f320cf4ac440a706918641d5da83e7f9630b2ae55ecac2fcca20a +size 203304 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YnriEa916FaU.flac b/Foundation/Audio_Grounding_AudioGrounding/YnriEa916FaU.flac new file mode 100644 index 0000000000000000000000000000000000000000..9a0bd4e41fe767f4da2e24ea6b535dfda0c8e7e9 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YnriEa916FaU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f765a78240f698f4fe7b8dc7201788cbbecdc5fa3e5ff7f2da82675ebc039e5d +size 230141 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yo7VIBi-2wg4.flac b/Foundation/Audio_Grounding_AudioGrounding/Yo7VIBi-2wg4.flac new file mode 100644 index 0000000000000000000000000000000000000000..5f71c9e5a2b83778d8d2c2f2eeb7bf543033e5fb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yo7VIBi-2wg4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c12fcd1e41e1ad4213f0bd751b49030cfa04e38ba7c18bde45f47b3a2596d9cf +size 231648 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YoHfoEcqtWgs.flac b/Foundation/Audio_Grounding_AudioGrounding/YoHfoEcqtWgs.flac new file mode 100644 index 0000000000000000000000000000000000000000..70b02e4df0d8284993024876451bf240ee136384 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YoHfoEcqtWgs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fc3d7bcb43daa2cb181130c65a2a2435fafb2d7187fcaf3f669ee3846eaf33a +size 216521 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YoN0IcZaHD_8.flac b/Foundation/Audio_Grounding_AudioGrounding/YoN0IcZaHD_8.flac new file mode 100644 index 0000000000000000000000000000000000000000..213f83f2d81052bb351d375d4ca5aa26ee478d29 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YoN0IcZaHD_8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69e4d3f633d8a9895e2199431817d23cc9d19ee0f89c47b1df9e0c4dff9b2bbb +size 217621 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YoNSWAH9mRxg.flac b/Foundation/Audio_Grounding_AudioGrounding/YoNSWAH9mRxg.flac new file mode 100644 index 0000000000000000000000000000000000000000..1546da0e378a219eacd2b88df6c6fe556eb6ba7d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YoNSWAH9mRxg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:036d310050c3f214e591b633a4ae42a276a88908cbf3c56d1f37293195a2e703 +size 223252 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YodJmGjFIhX8.flac b/Foundation/Audio_Grounding_AudioGrounding/YodJmGjFIhX8.flac new file mode 100644 index 0000000000000000000000000000000000000000..689179668d949795fa74f5160f93702cf8bd2772 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YodJmGjFIhX8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85258c5619ecafd06bdd0ae17c89d3913c65dac26c6d097519eb0af602fe0321 +size 252934 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YoqD-U-HYM6I.flac b/Foundation/Audio_Grounding_AudioGrounding/YoqD-U-HYM6I.flac new file mode 100644 index 0000000000000000000000000000000000000000..26060b870e89f3e1a7d85cff158de2fb3b8dbdad --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YoqD-U-HYM6I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d352c93b26d119989ba488371dddbd09e5fa6071ab4997565b7d6850dc735aa4 +size 140719 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YpWQeV08kYR0.flac b/Foundation/Audio_Grounding_AudioGrounding/YpWQeV08kYR0.flac new file mode 100644 index 0000000000000000000000000000000000000000..d287ade51a4f7a88002fcf4e4bbdda184e1c6bba --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YpWQeV08kYR0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d4cf4d5c10d5e1320734a78ff87800796187eb39042dea647d90c47abe02e2a +size 257246 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YpbiQmndSpoI.flac b/Foundation/Audio_Grounding_AudioGrounding/YpbiQmndSpoI.flac new file mode 100644 index 0000000000000000000000000000000000000000..c0cc7f81e0a61bf8b5586c26be0ccee4370d3938 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YpbiQmndSpoI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f4c370d00e14924d90b52f35064007e383ed63bb2dc5f4d28f32263bd041a4d +size 266153 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YpgzTXMgU1FM.flac b/Foundation/Audio_Grounding_AudioGrounding/YpgzTXMgU1FM.flac new file mode 100644 index 0000000000000000000000000000000000000000..ac842cab22f8ec33053df6b39d29da6a82c1f0a0 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YpgzTXMgU1FM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5737af14afc15ceb3a3b1403f02e67eab5ccda321794de331e56c8631fc4f160 +size 271172 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ypo-iqaVg9z8.flac b/Foundation/Audio_Grounding_AudioGrounding/Ypo-iqaVg9z8.flac new file mode 100644 index 0000000000000000000000000000000000000000..0200ace23e90dd05756f69910af4fa6fec4a6f7e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ypo-iqaVg9z8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6e44a205a443d6e2e707047f06d77c1a2d43cc822a3eb99bab9acbd06057422 +size 225658 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ypoo0WZI_ERw.flac b/Foundation/Audio_Grounding_AudioGrounding/Ypoo0WZI_ERw.flac new file mode 100644 index 0000000000000000000000000000000000000000..f08afa1ef5e699d5f52dfc6c9aa7960dc7df648e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ypoo0WZI_ERw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8b4ce1ba05baa285101d8ddd4d7eb7fb1400cf9bb857d58c802cdf9b834a5b68 +size 182873 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YptIksg9KEac.flac b/Foundation/Audio_Grounding_AudioGrounding/YptIksg9KEac.flac new file mode 100644 index 0000000000000000000000000000000000000000..5a4f7974cb667199c89854b37d4e1cb7805e01c3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YptIksg9KEac.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:afe6cfb4a9922f6d7159dbc2728667b8c5f1070af6c8b3ae44fa807efffed62b +size 197962 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YqGJKXoKTN4A.flac b/Foundation/Audio_Grounding_AudioGrounding/YqGJKXoKTN4A.flac new file mode 100644 index 0000000000000000000000000000000000000000..59bb1c0e7e4b30a495bf096f863b1a9ed9c73221 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YqGJKXoKTN4A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69997c29f58f70592ace7b2f264e30f160de9b01eb6445f914d27df29f1c3f89 +size 192503 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YqHDCNzaO208.flac b/Foundation/Audio_Grounding_AudioGrounding/YqHDCNzaO208.flac new file mode 100644 index 0000000000000000000000000000000000000000..59447e65ac13fb75f9af49b2a75fea2a95ddd8a1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YqHDCNzaO208.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81d00550309d453070d3e880f3cae358267cd0864b837978602dc25fac31b504 +size 225906 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YqOA4vLvHHn8.flac b/Foundation/Audio_Grounding_AudioGrounding/YqOA4vLvHHn8.flac new file mode 100644 index 0000000000000000000000000000000000000000..b4b0b90607dca7c5525c72d93ef005eaa6dca3b4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YqOA4vLvHHn8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc4b3da5a6ddbc2d2538a1f79136756de4f052722e8740cbbbfb37d631ab2511 +size 215891 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YqdRkvYNjkNs.flac b/Foundation/Audio_Grounding_AudioGrounding/YqdRkvYNjkNs.flac new file mode 100644 index 0000000000000000000000000000000000000000..47707363e3532096e436d43aafec45475da77362 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YqdRkvYNjkNs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88b09dd31eb1217181e98949e73da035152480aafc7dc8328e5bfa9603d7e346 +size 199221 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YrAaEi6xKe9A.flac b/Foundation/Audio_Grounding_AudioGrounding/YrAaEi6xKe9A.flac new file mode 100644 index 0000000000000000000000000000000000000000..8e769fd78d7facb7f70c633eb6bf5fb26b611d52 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YrAaEi6xKe9A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3160b74476f2fa5d8b200bbcf707235193c97f31e117b69478a466a7f17240f +size 241653 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YraPBAKyYIGE.flac b/Foundation/Audio_Grounding_AudioGrounding/YraPBAKyYIGE.flac new file mode 100644 index 0000000000000000000000000000000000000000..3ea5a282fe2f7f1af915c24e276887da23dfc578 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YraPBAKyYIGE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69097777f9219c83d8b2c5a5a472dfa82dc4c252fb760ff8fa07141832b7a704 +size 174506 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YriuGGhE1avc.flac b/Foundation/Audio_Grounding_AudioGrounding/YriuGGhE1avc.flac new file mode 100644 index 0000000000000000000000000000000000000000..c2849187eae1c714fe58355ded4cfc32bff99b13 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YriuGGhE1avc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1637f29f7401f917e2e4112eed46d53505aeaa8d64a42638cf8b73171340a06a +size 201359 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yrr5T3E4j7_Q.flac b/Foundation/Audio_Grounding_AudioGrounding/Yrr5T3E4j7_Q.flac new file mode 100644 index 0000000000000000000000000000000000000000..4d04e47d64a70e66028799d62e8a8a7a8a4d7dfb --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yrr5T3E4j7_Q.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d180357f4ed669275a4ab392a09a8cc4e4f734670cfd7456fe63ed62bbad4ef7 +size 189028 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YryTE7b5o4fE.flac b/Foundation/Audio_Grounding_AudioGrounding/YryTE7b5o4fE.flac new file mode 100644 index 0000000000000000000000000000000000000000..fbd13e2f3dbf4fa5f3e156529b5be50e8a4a3459 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YryTE7b5o4fE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:807f8619b0dc1405c4820a69ebbbec345ff7dac285c35f113eba4f1beea144bb +size 210776 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ys-1g0_KrLFI.flac b/Foundation/Audio_Grounding_AudioGrounding/Ys-1g0_KrLFI.flac new file mode 100644 index 0000000000000000000000000000000000000000..71a691c84c406ba7501b3853edc43508911a1c33 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ys-1g0_KrLFI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5307ed42dd7880f89a676d697e2c1e7ed439ed513a7272a0c027b8db8f7e0e0b +size 285908 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsM2tvbElLrI.flac b/Foundation/Audio_Grounding_AudioGrounding/YsM2tvbElLrI.flac new file mode 100644 index 0000000000000000000000000000000000000000..224c2b45c37adf59baa7bc3caae9e0bcfc19af3c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsM2tvbElLrI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fddae24a6023e28bb675f496070210170c66ae19709c0734c0644797d424f4ea +size 251944 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsNSmgP4P6Uc.flac b/Foundation/Audio_Grounding_AudioGrounding/YsNSmgP4P6Uc.flac new file mode 100644 index 0000000000000000000000000000000000000000..e3769c7ec06c1d6e82e18ce2ae048208fcc8ed52 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsNSmgP4P6Uc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11a512ab526cf61cdb9bd1c1b8ce12f069920422962d7d9505cf3e344884ee40 +size 223863 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsShpyu2l4YQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YsShpyu2l4YQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..64f4d05049e27e18564f19d8d628df3c7271d4a7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsShpyu2l4YQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9e1b3ec233d0da4cacade58a900a9b4310896222333488b150277b5cdfc58dd2 +size 205395 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsYj4hpDUZDQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YsYj4hpDUZDQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..ef423fe1632ed31aa727325776787a5b7f11a68a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsYj4hpDUZDQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:93acaf7635fda61ce423b680cb398d33a3bb9e25f05eb38ea02c4e3fc59073c0 +size 230393 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Ysd7x_Vssqlw.flac b/Foundation/Audio_Grounding_AudioGrounding/Ysd7x_Vssqlw.flac new file mode 100644 index 0000000000000000000000000000000000000000..b89a004b84396e6c52d23068d23c22b619db4d52 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Ysd7x_Vssqlw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f3663e38d9ca7047cc3091f4a2897699c81d48c7daa5ecef4d51a5137b6f9f9 +size 246108 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsiVH9ZLOZnQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YsiVH9ZLOZnQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..682ade23fdf4766624486c3688a5aa2f780317e7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsiVH9ZLOZnQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1065e582f251fde772315f5c119086f1dc43bd65a94f7c61d419f7fe6af93825 +size 253948 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsjlVMgdGSK0.flac b/Foundation/Audio_Grounding_AudioGrounding/YsjlVMgdGSK0.flac new file mode 100644 index 0000000000000000000000000000000000000000..b88974b9a83839b9b035749dbb46792b99181d28 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsjlVMgdGSK0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:da21f76b4173eab8806619857790ba5387929c411664e2af1ffb6a8e39584cf0 +size 270260 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YsofxkNWaP0s.flac b/Foundation/Audio_Grounding_AudioGrounding/YsofxkNWaP0s.flac new file mode 100644 index 0000000000000000000000000000000000000000..d293fd78d35eef9949202674390893c8c25a2485 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YsofxkNWaP0s.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:103e99149d9d6d4de59860bffb31e9c98eea7d676526975a7ec64d11129e878f +size 261518 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yt8vFu8WNVU0.flac b/Foundation/Audio_Grounding_AudioGrounding/Yt8vFu8WNVU0.flac new file mode 100644 index 0000000000000000000000000000000000000000..5025df41e4c5e6944ea68de31bf3198068a947fc --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yt8vFu8WNVU0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be0a657a9923b220aba57c4eadacc755b7ee3f37cd2c27c6a52a45b0a393d036 +size 192579 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YtDVADusiIoc.flac b/Foundation/Audio_Grounding_AudioGrounding/YtDVADusiIoc.flac new file mode 100644 index 0000000000000000000000000000000000000000..d245b710cd700f7ed92fe46e3f640a7da84bce25 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YtDVADusiIoc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82a170172e935c398783ef229ab2044eb20c9ac9579be3718dd45b70c06660dd +size 254562 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YtGWuKMJssrM.flac b/Foundation/Audio_Grounding_AudioGrounding/YtGWuKMJssrM.flac new file mode 100644 index 0000000000000000000000000000000000000000..0f72f4ba233870bfa877f112d1896e6fe6651352 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YtGWuKMJssrM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:188ac5e55b75940f08ee55a11b5bfad3d65781f712a77ad903c64d166bc3a784 +size 224876 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YtnUfrWBmrv8.flac b/Foundation/Audio_Grounding_AudioGrounding/YtnUfrWBmrv8.flac new file mode 100644 index 0000000000000000000000000000000000000000..ddb23ff82f127154f290ff999f8520a0bd731d34 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YtnUfrWBmrv8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:443a2432fd3a6858817da7ae3eb29a18bcd965f1cc82367b3b6262de82ab44ed +size 294286 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YtwgRUlPxFTk.flac b/Foundation/Audio_Grounding_AudioGrounding/YtwgRUlPxFTk.flac new file mode 100644 index 0000000000000000000000000000000000000000..cc7dea2e091f2a506d07c5a51ae4a6dbcefd4f6f --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YtwgRUlPxFTk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4cfbf1ee747068b4941e8253576004dd08e88ad393050a4fb94b85e8d3dca1a8 +size 201279 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YtznoShjZhyc.flac b/Foundation/Audio_Grounding_AudioGrounding/YtznoShjZhyc.flac new file mode 100644 index 0000000000000000000000000000000000000000..df49f64ac6ddf3fd9fa16faf28cdab8dbf75dc35 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YtznoShjZhyc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9835f43c08a689e88611269cdd58c3aa4b19476ec71c648d692456907d5d9bae +size 187865 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YuUV7Y4f0WCI.flac b/Foundation/Audio_Grounding_AudioGrounding/YuUV7Y4f0WCI.flac new file mode 100644 index 0000000000000000000000000000000000000000..beefaf3f0c7c561f732f57d5f521030cfc6e09a7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YuUV7Y4f0WCI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8df158b83d26f7d32e81d0799b827d258c35a79050ab095e98083c229380c9f +size 222457 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YuWPRNLnpy7Y.flac b/Foundation/Audio_Grounding_AudioGrounding/YuWPRNLnpy7Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..fd42e04c4909554f2b12552fb0643ffb1f765a67 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YuWPRNLnpy7Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:255b301203c809a591ec126a930e4972322182fecd3c4696947cc39c624acc61 +size 224600 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YufCPGoDcXLQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YufCPGoDcXLQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..d66939ac12378ca17234971a8da1fd556f60ec66 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YufCPGoDcXLQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a3975decf34ed65a511cb8fde337ce38330fe7815f2b7bd0bd984d9259799d30 +size 209437 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yulwx9Jx_hUg.flac b/Foundation/Audio_Grounding_AudioGrounding/Yulwx9Jx_hUg.flac new file mode 100644 index 0000000000000000000000000000000000000000..a7ac63765642bf2a06b9c6288c378d9357296c38 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yulwx9Jx_hUg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9c36954b15d464b7a8007870a316c14168400b543d9aec3c568ae4afb956b5c8 +size 227369 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yusasw640oaU.flac b/Foundation/Audio_Grounding_AudioGrounding/Yusasw640oaU.flac new file mode 100644 index 0000000000000000000000000000000000000000..80503b7fd54eb5d00f4fdef4c5f4ca8951111721 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yusasw640oaU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:007c61a3903b7b14887af9e0fc151a27581b54891b77d446fb60a6be85c39da5 +size 188721 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YvJ7JPEFhyLA.flac b/Foundation/Audio_Grounding_AudioGrounding/YvJ7JPEFhyLA.flac new file mode 100644 index 0000000000000000000000000000000000000000..237915416fe24f036e42b284be54584608a56954 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YvJ7JPEFhyLA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bc66f90f4243b59a93fae1e656662a7ea5adbe2506f948259b3a27beb4992a8c +size 255481 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YvNBZJqbiv0c.flac b/Foundation/Audio_Grounding_AudioGrounding/YvNBZJqbiv0c.flac new file mode 100644 index 0000000000000000000000000000000000000000..85c981e560955ba6448864b246d9762bb1a7c614 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YvNBZJqbiv0c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:44d80085cf52da498f4b1c29fe3b0cdf35ef0252189523bd30423d9f206d1195 +size 231877 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YvNNDoUbPQHE.flac b/Foundation/Audio_Grounding_AudioGrounding/YvNNDoUbPQHE.flac new file mode 100644 index 0000000000000000000000000000000000000000..7a3403323d19429db55902977559ac05ae14f35c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YvNNDoUbPQHE.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a946c02efbcabd81ddbe392d69d1dde35d011d84316c8c02ebd2e308d104cf88 +size 239201 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YvVhthZ45k3Y.flac b/Foundation/Audio_Grounding_AudioGrounding/YvVhthZ45k3Y.flac new file mode 100644 index 0000000000000000000000000000000000000000..1fd6f797e782954d0398bca1a4760c998e4c1a49 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YvVhthZ45k3Y.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:539c825473da34c01854a74b0e4a47c3e4bbdb1b3981e6e349ce82218dda7392 +size 247573 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YvhIYOLTyksQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YvhIYOLTyksQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..06388ae1606ec099c42af6919a51e726a92a5650 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YvhIYOLTyksQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:832036ea43cd5ac76096b8fc09dd25c233237329f6a4e2d3b030d3b8fe827c8e +size 222498 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YwF3poC0J20c.flac b/Foundation/Audio_Grounding_AudioGrounding/YwF3poC0J20c.flac new file mode 100644 index 0000000000000000000000000000000000000000..0390feb2eb6eed77b0dab6b1983c23325a27682a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YwF3poC0J20c.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9bdbb400469355cd40dbfb4c0ecbad7aec4274c90bce692b3441cba2471aee2 +size 283104 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YwI8Q_d2n68w.flac b/Foundation/Audio_Grounding_AudioGrounding/YwI8Q_d2n68w.flac new file mode 100644 index 0000000000000000000000000000000000000000..503e41494cb56454ae72211a4fa7d8bbad431827 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YwI8Q_d2n68w.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a36df954a922b8c8f109c754016158757ceb39362421a64f41649b5dbe979ace +size 234819 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YwVi5w_NU6CM.flac b/Foundation/Audio_Grounding_AudioGrounding/YwVi5w_NU6CM.flac new file mode 100644 index 0000000000000000000000000000000000000000..9475d8eaaf69e143ee8396adadb45d5bfb239a85 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YwVi5w_NU6CM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7ba7c64b36f6fa8d7cbd8ef8ad0bc3585ed57a1b7dcb8f02cd1a563da8e229f +size 284183 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YwqZ135_Ssz0.flac b/Foundation/Audio_Grounding_AudioGrounding/YwqZ135_Ssz0.flac new file mode 100644 index 0000000000000000000000000000000000000000..ae14af540d06e87a9888ddeea4db464621c4cabe --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YwqZ135_Ssz0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9ac4314748d367684370f57742ae000abde7690147839bfd252e396f3ad906f9 +size 190610 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YwwyfGO2J__4.flac b/Foundation/Audio_Grounding_AudioGrounding/YwwyfGO2J__4.flac new file mode 100644 index 0000000000000000000000000000000000000000..13ec39c473c3cd13f241d5d4aa65bba49b44f7ea --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YwwyfGO2J__4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8990416d7b32f047f862c9e8820d2144c2a392c01989fe2767282d3df666f001 +size 230823 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yx5cuQjOdM3E.flac b/Foundation/Audio_Grounding_AudioGrounding/Yx5cuQjOdM3E.flac new file mode 100644 index 0000000000000000000000000000000000000000..986501186065862928c3e048e703ae6696b5a555 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yx5cuQjOdM3E.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f192e1ba01288bfd71560f8df6cd4e9288e228e8b2801f5bfedf0fa4cc117d9 +size 216578 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YxIS6JwaGxuk.flac b/Foundation/Audio_Grounding_AudioGrounding/YxIS6JwaGxuk.flac new file mode 100644 index 0000000000000000000000000000000000000000..151b3193d76aa3cbc0564b0ea4cb1aadae70c351 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YxIS6JwaGxuk.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:90029089c8a6a975392b260d279ac1afcc3a98ee03afebb83b47f9a9093fe013 +size 259145 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YxUIdJ_aGp-A.flac b/Foundation/Audio_Grounding_AudioGrounding/YxUIdJ_aGp-A.flac new file mode 100644 index 0000000000000000000000000000000000000000..cf86ee10032b82fb9860745ab029aa716744dc0d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YxUIdJ_aGp-A.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8896daf9e5abfd1d0e5cb8466fa4463c01b33247b3afb20451dae5f6ec5c6192 +size 262456 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yxfaoyyzw2WU.flac b/Foundation/Audio_Grounding_AudioGrounding/Yxfaoyyzw2WU.flac new file mode 100644 index 0000000000000000000000000000000000000000..1e799dd71e2fa3e2579e524b9ccffe1758646ce1 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yxfaoyyzw2WU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:97006de79bc6cf040a4d14da41c86153c1812af42cfa3c027c3111ae1862ca5d +size 302669 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yxn4AgQ3Ekrc.flac b/Foundation/Audio_Grounding_AudioGrounding/Yxn4AgQ3Ekrc.flac new file mode 100644 index 0000000000000000000000000000000000000000..f172eb2f2f4db174a71a659e62fb2c2825e0d5ca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yxn4AgQ3Ekrc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:376afa05007c62f76d1865aaeb588949acd15afd392dabcef2c4c0341ce6240f +size 270966 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YxnVzaYbBspo.flac b/Foundation/Audio_Grounding_AudioGrounding/YxnVzaYbBspo.flac new file mode 100644 index 0000000000000000000000000000000000000000..f88a5137eda187673f9b9f6d37ea1987044b2afa --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YxnVzaYbBspo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5c9de8d13300fc8bf93cbd83cbbc3525bd7059f40fc2643c5b8fbd497a6c3dc +size 152828 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yxx3yKzrSOBI.flac b/Foundation/Audio_Grounding_AudioGrounding/Yxx3yKzrSOBI.flac new file mode 100644 index 0000000000000000000000000000000000000000..c5417f20a68ec06a0ea1c476dff80c16d18702e3 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yxx3yKzrSOBI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95813993723e43ca269d4ae119ee8fe459a8f557b26c4862665be8106235837b +size 261558 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YxxPndHjWTCA.flac b/Foundation/Audio_Grounding_AudioGrounding/YxxPndHjWTCA.flac new file mode 100644 index 0000000000000000000000000000000000000000..6b4c0ee26b7c903c5d6f456b12f897074ae7289b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YxxPndHjWTCA.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a26487c14ca5293bb0562a45d96f48a3e78e2472654ea62abd72194545c23f33 +size 255685 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yxz6IFHbr_Ac.flac b/Foundation/Audio_Grounding_AudioGrounding/Yxz6IFHbr_Ac.flac new file mode 100644 index 0000000000000000000000000000000000000000..76619f4d47f8f0fa8e9ce36790568403270199fe --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yxz6IFHbr_Ac.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f03f922581b5898a7eb7da83b398dff159051d615cc750a7d90f5ee0f3343768 +size 264459 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy0PVvrn_uGU.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy0PVvrn_uGU.flac new file mode 100644 index 0000000000000000000000000000000000000000..7eeb901486cccee6189bbcef22fb72f3082a3637 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy0PVvrn_uGU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f48a21221bd6dea26c78482215237f33d782540069031d08b3901daead437815 +size 235513 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy1saVTXsKwc.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy1saVTXsKwc.flac new file mode 100644 index 0000000000000000000000000000000000000000..1ceee50bb507756a0ff0860b76c0ffee504b747c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy1saVTXsKwc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b195addc8633e736fbd380fe9ed11ca99c377d315ab2bafd1e166cfc586ef671 +size 242943 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy2ZBGpgbhHM.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy2ZBGpgbhHM.flac new file mode 100644 index 0000000000000000000000000000000000000000..b5d93105fa4542995f8be58584a8b9dcccc6c749 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy2ZBGpgbhHM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0051b48e8f3f3825d47e18f2bdca691a35edb8edf30a746c19870861a14496ee +size 155600 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy3alb754Sis.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy3alb754Sis.flac new file mode 100644 index 0000000000000000000000000000000000000000..ebb5e8375042f6fa7682fc3d7f139d48540528ca --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy3alb754Sis.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ef4c77db6b86c7f3d9587c703440d0b139118c98fdfd649f4e27a4eb5eb547e9 +size 231250 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy6wsRU2aNx4.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy6wsRU2aNx4.flac new file mode 100644 index 0000000000000000000000000000000000000000..2020cd794eafec16f7c7fafae1af5985a4efc036 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy6wsRU2aNx4.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9179133b2685809f744204a7b18484cfb35b79dbbed8969691fcbdbcfd7b0b29 +size 226844 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yy8dSeubCN_I.flac b/Foundation/Audio_Grounding_AudioGrounding/Yy8dSeubCN_I.flac new file mode 100644 index 0000000000000000000000000000000000000000..24fd17e098b0ad70c7155187bf19ff7e8c2898dd --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yy8dSeubCN_I.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fed8cfd90b71fa96d00166f69cb812ca55cd38cce284018d82ae3851b910253 +size 257531 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyHq2AHuV2BU.flac b/Foundation/Audio_Grounding_AudioGrounding/YyHq2AHuV2BU.flac new file mode 100644 index 0000000000000000000000000000000000000000..1b3b3d57e6e7a49ed163768ba116cfa23718c56a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyHq2AHuV2BU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f750e55d7daa58f4b8a260e6c1158af7bbfe9d1393f68e7365ff80b373491128 +size 164234 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyJ-KUG1m2CY.flac b/Foundation/Audio_Grounding_AudioGrounding/YyJ-KUG1m2CY.flac new file mode 100644 index 0000000000000000000000000000000000000000..8957ccc3c3874ecd57f770720472ee9c7264bc1e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyJ-KUG1m2CY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e5180c44ac8a6d60455a7ba1192c74ce0d7269ed0b40d18e291cb9865d8eb06 +size 145306 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyK0XUXQULZs.flac b/Foundation/Audio_Grounding_AudioGrounding/YyK0XUXQULZs.flac new file mode 100644 index 0000000000000000000000000000000000000000..881663b86a3d67b7485fa803ab34160133d76924 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyK0XUXQULZs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f38bafde7883e8641c7c265167a1e0788d9ad876c8bafbf5f22bb95a1cf70c0 +size 221509 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyNxFTGu3TBg.flac b/Foundation/Audio_Grounding_AudioGrounding/YyNxFTGu3TBg.flac new file mode 100644 index 0000000000000000000000000000000000000000..58417e2bf53563ce2ac614ddc2b7af967a40995b --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyNxFTGu3TBg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39834273f0de8d593e9ef13cc86554ed566dc6085b84b0a9198fff641aee3d9d +size 212778 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyRUZQSU4EzQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YyRUZQSU4EzQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..157b46e5f956d1d0faeb48796e8054c9e51f2d5a --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyRUZQSU4EzQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4aa2e8c2a7ebbbc7ebf05ca0302ffa767ef12099a4734257c4f34d6efaeb2ee9 +size 214376 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyXnZjfMfzdI.flac b/Foundation/Audio_Grounding_AudioGrounding/YyXnZjfMfzdI.flac new file mode 100644 index 0000000000000000000000000000000000000000..fcfab24e3281ecec5136f80b01f30be76cc832f8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyXnZjfMfzdI.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00e7f56ef758e50515146668d6da2383ada34e5827fdf752716852b847dd99d4 +size 215991 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YydTFISfwpRo.flac b/Foundation/Audio_Grounding_AudioGrounding/YydTFISfwpRo.flac new file mode 100644 index 0000000000000000000000000000000000000000..81d3a3e3ba17182edca241925b877cfc903d967e --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YydTFISfwpRo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5273fe4035170ab7524a23ad57a4231760a32886cd718d7699ea6d6c39442d00 +size 222600 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YymXRSkBSEJs.flac b/Foundation/Audio_Grounding_AudioGrounding/YymXRSkBSEJs.flac new file mode 100644 index 0000000000000000000000000000000000000000..57dc3143adb91a82da395a7bdd701e3183769c74 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YymXRSkBSEJs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b0f438ef4754c8a253eb2689d0edc12a4492030af2f96068fb7db2bdf476b82 +size 194041 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyrP1c4KzHFo.flac b/Foundation/Audio_Grounding_AudioGrounding/YyrP1c4KzHFo.flac new file mode 100644 index 0000000000000000000000000000000000000000..e4930a842d06d533bd9e1880feb2762c81e57488 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyrP1c4KzHFo.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:09ccae66862f11e7faf693add6f3d7732da437b663ea442265d64b4c7bc50a19 +size 265777 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yysdfo6Am1jc.flac b/Foundation/Audio_Grounding_AudioGrounding/Yysdfo6Am1jc.flac new file mode 100644 index 0000000000000000000000000000000000000000..6c57f992d1c343437c08c12a02a6124d4b549f29 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yysdfo6Am1jc.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9de12730fd788f24257e80424c2800ff1d41f51b04285f2568403cb987b0ec55 +size 115682 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YyyGk2dAOjZM.flac b/Foundation/Audio_Grounding_AudioGrounding/YyyGk2dAOjZM.flac new file mode 100644 index 0000000000000000000000000000000000000000..a4eda640e229a371ddb6024292d92ee089ac5047 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YyyGk2dAOjZM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:23f986869ff5cc1165b59564036f471edb7034d896fbab0aca2d236b48ee8331 +size 147535 diff --git a/Foundation/Audio_Grounding_AudioGrounding/Yz1dShwq36TY.flac b/Foundation/Audio_Grounding_AudioGrounding/Yz1dShwq36TY.flac new file mode 100644 index 0000000000000000000000000000000000000000..406d446a4c4536e7a8b657baaa3beeaa19bbc165 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/Yz1dShwq36TY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a5f51ff7f8737ecf20ae03b028f2aaa31f2d53cb60b81b1bc8cedd45372aaa2 +size 246844 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzB4kz0vEsTQ.flac b/Foundation/Audio_Grounding_AudioGrounding/YzB4kz0vEsTQ.flac new file mode 100644 index 0000000000000000000000000000000000000000..cf6a3bf9ed0ad63efe0f2e36a08bfbd0579f5072 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzB4kz0vEsTQ.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25f2bca6c46778f8a9e2052e0332218cee4844c3d22417f9385aa1d0064d567e +size 212012 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzG0dfs6IAl0.flac b/Foundation/Audio_Grounding_AudioGrounding/YzG0dfs6IAl0.flac new file mode 100644 index 0000000000000000000000000000000000000000..91a80b8476219fbf0991a72c3e6ca465a39f2e85 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzG0dfs6IAl0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d8a9f253a99414056b3f9d95fc82cf169c8d5d068276c8070f47a8198b8c2ac3 +size 249963 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzM4rnZcNY8U.flac b/Foundation/Audio_Grounding_AudioGrounding/YzM4rnZcNY8U.flac new file mode 100644 index 0000000000000000000000000000000000000000..bcbc611986a7ba5d826e855d5f13a0526617ce94 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzM4rnZcNY8U.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a7c3c39fddec384d662992ef286a58b9df9462e643be0fde8f775cb81fab22d +size 217016 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzV0gxh-yoqs.flac b/Foundation/Audio_Grounding_AudioGrounding/YzV0gxh-yoqs.flac new file mode 100644 index 0000000000000000000000000000000000000000..aa34139c2f4d7b2170f6fe2c5e15146cf29e3ff7 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzV0gxh-yoqs.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:834c549a79e9f69aa219ce35272917fcda98e06487a7459966201bea9ba1a1bd +size 218381 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzX0OMIdZERU.flac b/Foundation/Audio_Grounding_AudioGrounding/YzX0OMIdZERU.flac new file mode 100644 index 0000000000000000000000000000000000000000..de76e48e8f64904d80f525da8b2681cc1b65051c --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzX0OMIdZERU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6f1334f609235a445c7dd635fb32f95eb00941ebbc27deea1bf41bbc0221d3ba +size 275205 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzbeBikc3Qt8.flac b/Foundation/Audio_Grounding_AudioGrounding/YzbeBikc3Qt8.flac new file mode 100644 index 0000000000000000000000000000000000000000..160d3073f97c05b15acaeec2f9b949bf70cced48 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzbeBikc3Qt8.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5deb62920311909e1518c8a84f2128548d987d5d0903465b90b7c635bbdd255d +size 273812 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzguSB2RIKpw.flac b/Foundation/Audio_Grounding_AudioGrounding/YzguSB2RIKpw.flac new file mode 100644 index 0000000000000000000000000000000000000000..d61ce889c2f698c0ccd6e1c054e356a668afae84 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzguSB2RIKpw.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:803c314e4680a0598bd7676036cb353565578e3059e0e5a5b954658706521992 +size 249626 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YziAl_IGwMpM.flac b/Foundation/Audio_Grounding_AudioGrounding/YziAl_IGwMpM.flac new file mode 100644 index 0000000000000000000000000000000000000000..8cb99a2a7c7c75d8acf59986c5408ee3439f9be5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YziAl_IGwMpM.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b801e8d0a9bfc36280d7330a557149b91f5cc645ec417a73de226503a32274a5 +size 245323 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YziQzaDcMzJg.flac b/Foundation/Audio_Grounding_AudioGrounding/YziQzaDcMzJg.flac new file mode 100644 index 0000000000000000000000000000000000000000..645fe7431956325542e9619c1bb5cbc4b44e6fa4 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YziQzaDcMzJg.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:410a34692d7db6909a3b33cf276f578144626dc56e4fe9c28ac43619ef672fff +size 233012 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzlJ8jLkCRUU.flac b/Foundation/Audio_Grounding_AudioGrounding/YzlJ8jLkCRUU.flac new file mode 100644 index 0000000000000000000000000000000000000000..16c720e169bac55559ca1c77c1f7c2ce3fd516c8 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzlJ8jLkCRUU.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fd6e31e2b4290f9ed11bfd0d5ee6162d54f6e8311a3ba1577306869f15dc7df +size 226529 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzlL_aOhD_X0.flac b/Foundation/Audio_Grounding_AudioGrounding/YzlL_aOhD_X0.flac new file mode 100644 index 0000000000000000000000000000000000000000..3a1d07a6956c45730022ba55fc997a94c48d05b5 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzlL_aOhD_X0.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2cd08bb15b2afefdafeed845905b4b9b6bf6f925cee289b52a36ec5d41b85de0 +size 231591 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzpsOHpUwUfY.flac b/Foundation/Audio_Grounding_AudioGrounding/YzpsOHpUwUfY.flac new file mode 100644 index 0000000000000000000000000000000000000000..171f3bf5620988a7acc202c85732c5d9d6788570 --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzpsOHpUwUfY.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a522142adf260ec41969bebe9226fdafdf8659e4badc1f12f249e5cfd494144 +size 239123 diff --git a/Foundation/Audio_Grounding_AudioGrounding/YzsoRpeRkN34.flac b/Foundation/Audio_Grounding_AudioGrounding/YzsoRpeRkN34.flac new file mode 100644 index 0000000000000000000000000000000000000000..d471c55820f3ceb7194280be5feff28a18b6633d --- /dev/null +++ b/Foundation/Audio_Grounding_AudioGrounding/YzsoRpeRkN34.flac @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:495535b250b95053f8327fecb7bb79411bfe3a14e4ef0d3af201b0207b7f0dbe +size 314844 diff --git a/Foundation/Foundation_meta.json b/Foundation/Foundation_meta.json new file mode 100644 index 0000000000000000000000000000000000000000..5edfcdc0d4bc6446652a7112e6bcba50300bf165 --- /dev/null +++ b/Foundation/Foundation_meta.json @@ -0,0 +1,291110 @@ +[ + { + "path": "2902-9008-0013.flac", + "question": "Highlight when 'hate' is spoken.", + "choice_a": "[7.67, 8.05]", + "choice_b": "[1.03, 1.53]", + "choice_c": "[3.07, 3.27]", + "choice_d": "[7.02, 7.21]", + "answer_gt": "[7.67, 8.05]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 0 + }, + { + "path": "4515-11057-0078.flac", + "question": "What is the duration of 'him' being mentioned in the conversation?", + "choice_a": "[1.77, 2.03]", + "choice_b": "[0.34, 0.46]", + "choice_c": "[2.54, 2.66]", + "choice_d": "[2.34, 2.54]", + "answer_gt": "[1.77, 2.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 1 + }, + { + "path": "3528-168669-0004.flac", + "question": "Pinpoint the exact moment 'duties' is repeated in the audio.", + "choice_a": "[2.63, 3.11]", + "choice_b": "[4.24, 4.31]", + "choice_c": "[2.45, 2.52]", + "choice_d": "[1.01, 1.17]", + "answer_gt": "[2.63, 3.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 2 + }, + { + "path": "84-121123-0005.flac", + "question": "Establish when 'explanation' is first mentioned in the clip.", + "choice_a": "[1.79, 1.89]", + "choice_b": "[6.08, 6.87]", + "choice_c": "[2.48, 12.6]", + "choice_d": "[3.02, 3.28]", + "answer_gt": "[6.08, 6.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 3 + }, + { + "path": "3915-98647-0034.flac", + "question": "Determine the time when 'another' happens within the clip.", + "choice_a": "[1.17, 1.52]", + "choice_b": "[2.74, 3.32]", + "choice_c": "[0.49, 0.81]", + "choice_d": "[3.56, 3.69]", + "answer_gt": "[0.49, 0.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 4 + }, + { + "path": "2412-153947-0015.flac", + "question": "Find the duration when 'fairly' occurs consecutively in the audio.", + "choice_a": "[7.88, 8.21]", + "choice_b": "[0.80, 1.15]", + "choice_c": "[6.38, 6.43]", + "choice_d": "[8.82, 9.13]", + "answer_gt": "[0.80, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 5 + }, + { + "path": "3331-159605-0013.flac", + "question": "Highlight the time when 'at' is mentioned.", + "choice_a": "[0.54, 0.65]", + "choice_b": "[2.03, 2.29]", + "choice_c": "[7.57, 8.17]", + "choice_d": "[6.79, 6.96]", + "answer_gt": "[0.54, 0.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 6 + }, + { + "path": "5442-41168-0001.flac", + "question": "Highlight the time when 'his' is mentioned.", + "choice_a": "[3.19, 3.35]", + "choice_b": "[3.35, 3.68]", + "choice_c": "[4.33, 4.83]", + "choice_d": "[1.19, 1.26]", + "answer_gt": "[3.19, 3.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 7 + }, + { + "path": "8288-274150-0008.flac", + "question": "Listen to this audio clip; identify the time segment when 'utter' is spoken.", + "choice_a": "[2.36, 2.84]", + "choice_b": "[0.64, 1.13]", + "choice_c": "[4.57, 4.79]", + "choice_d": "[1.79, 2.09]", + "answer_gt": "[4.57, 4.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 8 + }, + { + "path": "3915-57461-0019.flac", + "question": "Locate the segment where 'learned' appears most frequently.", + "choice_a": "[4.89, 5.05]", + "choice_b": "[2.73, 2.99]", + "choice_c": "[5.58, 6.06]", + "choice_d": "[1.18, 1.64]", + "answer_gt": "[2.73, 2.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 9 + }, + { + "path": "3915-98647-0021.flac", + "question": "Spot the exact moment 'small' appears in the audio.", + "choice_a": "[7.05, 7.52]", + "choice_b": "[1.71, 2.06]", + "choice_c": "[9.82, 10.1]", + "choice_d": "[7.62, 7.99]", + "answer_gt": "[7.62, 7.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 10 + }, + { + "path": "672-122797-0056.flac", + "question": "When is 'what' first heard in the recording?", + "choice_a": "[2.68, 3.00]", + "choice_b": "[0.37, 0.59]", + "choice_c": "[0.66, 1.01]", + "choice_d": "[3.50, 4.13]", + "answer_gt": "[2.68, 3.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 11 + }, + { + "path": "908-157963-0007_0.flac", + "question": "Note the time interval that includes the mention of 'head'.", + "choice_a": "[1.14, 1.65]", + "choice_b": "[6.76, 6.94]", + "choice_c": "[6.52, 16.7]", + "choice_d": "[6.92, 17.3]", + "answer_gt": "[6.92, 17.3]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 12 + }, + { + "path": "4852-28319-0005.flac", + "question": "Establish when 'to' is first mentioned in the clip.", + "choice_a": "[4.52, 4.63]", + "choice_b": "[0.93, 1.57]", + "choice_c": "[6.30, 6.35]", + "choice_d": "[1.57, 1.81]", + "answer_gt": "[4.52, 4.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 13 + }, + { + "path": "4294-35475-0015.flac", + "question": "Highlight when 'wall' is spoken.", + "choice_a": "[0.45, 0.80]", + "choice_b": "[0.30, 0.45]", + "choice_c": "[3.18, 3.64]", + "choice_d": "[2.47, 2.88]", + "answer_gt": "[2.47, 2.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 14 + }, + { + "path": "174-168635-0021.flac", + "question": "At what time interval does 'strengthened' occur in the audio clip?", + "choice_a": "[0.45, 0.68]", + "choice_b": "[2.12, 2.48]", + "choice_c": "[2.48, 3.31]", + "choice_d": "[1.30, 1.62]", + "answer_gt": "[2.48, 3.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 15 + }, + { + "path": "251-137823-0015.flac", + "question": "When does the audio feature the sound of 'help'?", + "choice_a": "[1.04, 1.26]", + "choice_b": "[0.13, 0.33]", + "choice_c": "[0.76, 1.04]", + "choice_d": "[0.50, 0.76]", + "answer_gt": "[1.04, 1.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 16 + }, + { + "path": "2609-169640-0004.flac", + "question": "Ascertain the point at which 'barely' enters the discussion.", + "choice_a": "[9.59, 9.85]", + "choice_b": "[5.08, 5.19]", + "choice_c": "[7.45, 8.03]", + "choice_d": "[8.06, 8.27]", + "answer_gt": "[9.59, 9.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 17 + }, + { + "path": "7975-280085-0003.flac", + "question": "Find out when the speaker refers to 'took' during the audio.", + "choice_a": "[1.03, 1.25]", + "choice_b": "[0.74, 1.03]", + "choice_c": "[2.48, 2.73]", + "choice_d": "[2.73, 2.86]", + "answer_gt": "[2.48, 2.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 18 + }, + { + "path": "8455-210777-0017.flac", + "question": "Note the time when 'on' is first mentioned in the audio.", + "choice_a": "[1.62, 1.76]", + "choice_b": "[1.81, 2.09]", + "choice_c": "[4.00, 4.29]", + "choice_d": "[2.09, 2.15]", + "answer_gt": "[1.62, 1.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 19 + }, + { + "path": "7902-96592-0033.flac", + "question": "Listen for when 'you' is emphasized by the narrator.", + "choice_a": "[0.47, 0.85]", + "choice_b": "[2.16, 2.33]", + "choice_c": "[2.36, 2.79]", + "choice_d": "[3.23, 3.33]", + "answer_gt": "[2.16, 2.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 20 + }, + { + "path": "2033-164915-0014.flac", + "question": "When is 'forwards' first mentioned in the clip?", + "choice_a": "[1.08, 1.94]", + "choice_b": "[3.46, 3.82]", + "choice_c": "[0.37, 0.57]", + "choice_d": "[5.40, 5.91]", + "answer_gt": "[1.08, 1.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 21 + }, + { + "path": "8461-281231-0014.flac", + "question": "Ascertain the exact time 'thy' is mentioned in the audio.", + "choice_a": "[0.90, 1.37]", + "choice_b": "[0.71, 0.90]", + "choice_c": "[2.32, 2.81]", + "choice_d": "[1.83, 2.21]", + "answer_gt": "[0.71, 0.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 22 + }, + { + "path": "533-131556-0023.flac", + "question": "Locate the segment where 'the' is faintly heard.", + "choice_a": "[2.32, 2.41]", + "choice_b": "[3.44, 3.51]", + "choice_c": "[1.68, 1.82]", + "choice_d": "[0.30, 0.42]", + "answer_gt": "[2.32, 2.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 23 + }, + { + "path": "4852-28319-0014.flac", + "question": "Note the time when 'idea' is first mentioned in the audio.", + "choice_a": "[1.08, 1.44]", + "choice_b": "[1.47, 1.89]", + "choice_c": "[1.89, 2.00]", + "choice_d": "[0.51, 0.65]", + "answer_gt": "[1.08, 1.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 24 + }, + { + "path": "1630-141772-0014.flac", + "question": "What is the timestamp for the initial occurrence of 'with'?", + "choice_a": "[1.86, 2.19]", + "choice_b": "[3.20, 3.58]", + "choice_c": "[0.79, 1.16]", + "choice_d": "[1.70, 1.86]", + "answer_gt": "[1.70, 1.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 25 + }, + { + "path": "8131-117016-0058.flac", + "question": "Locate the segment where 'training' appears most frequently.", + "choice_a": "[3.23, 3.64]", + "choice_b": "[3.64, 3.70]", + "choice_c": "[1.40, 2.29]", + "choice_d": "[2.48, 2.67]", + "answer_gt": "[3.23, 3.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 26 + }, + { + "path": "3005-163399-0011.flac", + "question": "When does the speaker mention 'i'?", + "choice_a": "[1.56, 2.03]", + "choice_b": "[2.21, 2.36]", + "choice_c": "[2.03, 2.21]", + "choice_d": "[1.19, 1.56]", + "answer_gt": "[2.21, 2.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 27 + }, + { + "path": "4153-185072-0001.flac", + "question": "Recognize the time when 'his' is most clearly articulated.", + "choice_a": "[4.22, 4.66]", + "choice_b": "[9.14, 9.32]", + "choice_c": "[3.95, 4.13]", + "choice_d": "[5.20, 5.44]", + "answer_gt": "[9.14, 9.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 28 + }, + { + "path": "8254-84205-0053.flac", + "question": "At what time interval does 'made' occur in the audio clip?", + "choice_a": "[2.02, 2.30]", + "choice_b": "[0.86, 0.95]", + "choice_c": "[0.95, 1.24]", + "choice_d": "[2.40, 2.53]", + "answer_gt": "[2.02, 2.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 29 + }, + { + "path": "1988-24833-0025.flac", + "question": "Highlight the time when 'a' is mentioned.", + "choice_a": "[1.46, 1.51]", + "choice_b": "[1.02, 1.40]", + "choice_c": "[0.21, 0.31]", + "choice_d": "[1.51, 1.73]", + "answer_gt": "[1.46, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 30 + }, + { + "path": "1255-90407-0023.flac", + "question": "Locate the segment where 'a' appears most frequently.", + "choice_a": "[6.88, 7.01]", + "choice_b": "[6.08, 6.28]", + "choice_c": "[3.82, 3.98]", + "choice_d": "[2.55, 2.59]", + "answer_gt": "[2.55, 2.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 31 + }, + { + "path": "5484-24317-0025.flac", + "question": "At what interval does 'confirmed' appear in the audio?", + "choice_a": "[2.75, 2.83]", + "choice_b": "[6.79, 7.29]", + "choice_c": "[5.81, 6.32]", + "choice_d": "[2.20, 2.44]", + "answer_gt": "[6.79, 7.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 32 + }, + { + "path": "3331-159605-0040.flac", + "question": "At what interval does 'thought' appear in the audio?", + "choice_a": "[0.33, 0.37]", + "choice_b": "[0.61, 0.79]", + "choice_c": "[2.36, 2.72]", + "choice_d": "[0.37, 0.61]", + "answer_gt": "[2.36, 2.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 33 + }, + { + "path": "2609-169640-0011.flac", + "question": "At what time interval does 'movements' occur in the audio clip?", + "choice_a": "[4.48, 4.64]", + "choice_b": "[3.66, 4.21]", + "choice_c": "[4.64, 4.86]", + "choice_d": "[0.57, 0.94]", + "answer_gt": "[3.66, 4.21]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 34 + }, + { + "path": "3997-180297-0011.flac", + "question": "Identify the moment 'enough' is mentioned in the audio.", + "choice_a": "[1.03, 1.59]", + "choice_b": "[2.48, 2.65]", + "choice_c": "[3.04, 3.35]", + "choice_d": "[3.35, 3.45]", + "answer_gt": "[3.04, 3.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 35 + }, + { + "path": "1995-1826-0023.flac", + "question": "Spot the exact moment 'on' appears in the audio.", + "choice_a": "[3.32, 4.18]", + "choice_b": "[2.04, 2.16]", + "choice_c": "[1.62, 2.04]", + "choice_d": "[1.37, 1.56]", + "answer_gt": "[1.37, 1.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 36 + }, + { + "path": "8297-275156-0007.flac", + "question": "Identify when 'the' is heard in the audio.", + "choice_a": "[7.85, 8.25]", + "choice_b": "[4.64, 4.76]", + "choice_c": "[2.30, 3.10]", + "choice_d": "[0.89, 0.99]", + "answer_gt": "[0.89, 0.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 37 + }, + { + "path": "4852-28319-0012.flac", + "question": "Determine the precise moment 'the' in the audio's pace.", + "choice_a": "[2.21, 2.27]", + "choice_b": "[1.29, 1.52]", + "choice_c": "[0.63, 1.18]", + "choice_d": "[0.18, 0.63]", + "answer_gt": "[2.21, 2.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 38 + }, + { + "path": "1272-141231-0014.flac", + "question": "Detect the point at which 'a' is spoken.", + "choice_a": "[0.97, 1.22]", + "choice_b": "[4.97, 5.06]", + "choice_c": "[0.84, 0.97]", + "choice_d": "[5.33, 5.40]", + "answer_gt": "[5.33, 5.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 39 + }, + { + "path": "8254-115543-0041.flac", + "question": "When in the audio does 'were' coincide with a specific event?", + "choice_a": "[4.08, 4.38]", + "choice_b": "[3.47, 3.61]", + "choice_c": "[0.93, 0.99]", + "choice_d": "[1.32, 1.51]", + "answer_gt": "[0.93, 0.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 40 + }, + { + "path": "672-122797-0048.flac", + "question": "Pinpoint the exact moment 'if' appears.", + "choice_a": "[3.59, 3.96]", + "choice_b": "[0.48, 0.87]", + "choice_c": "[2.58, 2.79]", + "choice_d": "[2.45, 2.58]", + "answer_gt": "[0.48, 0.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 41 + }, + { + "path": "6123-59186-0036.flac", + "question": "What is the timestamp for when 'always' occurs?", + "choice_a": "[0.81, 1.25]", + "choice_b": "[2.51, 2.76]", + "choice_c": "[1.25, 1.67]", + "choice_d": "[0.74, 0.81]", + "answer_gt": "[0.81, 1.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 42 + }, + { + "path": "422-122949-0003.flac", + "question": "In the audio track, when is 'designations' first noticed?", + "choice_a": "[1.86, 2.70]", + "choice_b": "[3.83, 3.91]", + "choice_c": "[4.80, 15.3]", + "choice_d": "[6.45, 6.75]", + "answer_gt": "[1.86, 2.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 43 + }, + { + "path": "5764-299665-0016.flac", + "question": "Ascertain the point at which 'the' enters the discussion.", + "choice_a": "[4.16, 4.36]", + "choice_b": "[3.79, 4.16]", + "choice_c": "[3.15, 3.68]", + "choice_d": "[1.87, 1.99]", + "answer_gt": "[1.87, 1.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 44 + }, + { + "path": "6432-63722-0057.flac", + "question": "Note the time interval that includes the mention of 'holds'.", + "choice_a": "[0.32, 0.78]", + "choice_b": "[0.89, 1.18]", + "choice_c": "[1.18, 1.45]", + "choice_d": "[0.78, 0.89]", + "answer_gt": "[1.18, 1.45]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 45 + }, + { + "path": "4507-16021-0050.flac", + "question": "Note the time when 'class' is first mentioned in the audio.", + "choice_a": "[2.02, 2.47]", + "choice_b": "[0.77, 1.18]", + "choice_c": "[2.60, 2.81]", + "choice_d": "[1.33, 1.55]", + "answer_gt": "[2.02, 2.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 46 + }, + { + "path": "4570-14911-0010.flac", + "question": "Locate the part of the audio where 'mild' is expressed.", + "choice_a": "[6.71, 7.14]", + "choice_b": "[9.27, 9.80]", + "choice_c": "[0.62, 1.24]", + "choice_d": "[9.17, 9.27]", + "answer_gt": "[9.27, 9.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 47 + }, + { + "path": "2086-149220-0015.flac", + "question": "When does the audio feature the sound of 'have'?", + "choice_a": "[0.97, 1.11]", + "choice_b": "[0.55, 0.70]", + "choice_c": "[6.76, 7.14]", + "choice_d": "[1.42, 1.83]", + "answer_gt": "[0.55, 0.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 48 + }, + { + "path": "8254-115543-0011.flac", + "question": "When is 'was' first mentioned in the clip?", + "choice_a": "[4.59, 4.73]", + "choice_b": "[2.99, 3.10]", + "choice_c": "[3.57, 4.10]", + "choice_d": "[0.24, 0.49]", + "answer_gt": "[2.99, 3.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 49 + }, + { + "path": "7976-105575-0000.flac", + "question": "Locate the part of the audio where 'was' is expressed.", + "choice_a": "[6.26, 6.62]", + "choice_b": "[5.48, 5.65]", + "choice_c": "[5.65, 6.00]", + "choice_d": "[0.72, 0.86]", + "answer_gt": "[0.72, 0.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 50 + }, + { + "path": "652-130726-0013.flac", + "question": "Pinpoint the exact moment 'this' is repeated in the audio.", + "choice_a": "[4.33, 4.63]", + "choice_b": "[0.67, 0.78]", + "choice_c": "[4.21, 4.33]", + "choice_d": "[0.52, 0.67]", + "answer_gt": "[0.67, 0.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 51 + }, + { + "path": "4852-28330-0019.flac", + "question": "Listen for the occurrence of 'time' in the recording.", + "choice_a": "[2.35, 2.90]", + "choice_b": "[3.83, 4.22]", + "choice_c": "[2.99, 3.33]", + "choice_d": "[1.77, 1.83]", + "answer_gt": "[3.83, 4.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 52 + }, + { + "path": "4970-29095-0011.flac", + "question": "Listen for when 'thee' is emphasized by the narrator.", + "choice_a": "[0.56, 0.72]", + "choice_b": "[0.56, 0.72]", + "choice_c": "[2.16, 2.28]", + "choice_d": "[2.28, 2.60]", + "answer_gt": "[0.56, 0.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 53 + }, + { + "path": "1272-135031-0015.flac", + "question": "Recognize the time when 'largest' is most clearly articulated.", + "choice_a": "[0.62, 0.91]", + "choice_b": "[0.55, 0.62]", + "choice_c": "[3.79, 4.55]", + "choice_d": "[1.71, 1.85]", + "answer_gt": "[3.79, 4.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 54 + }, + { + "path": "1585-131718-0014.flac", + "question": "Detect the point at which 'the' is spoken.", + "choice_a": "[1.69, 2.05]", + "choice_b": "[1.15, 1.47]", + "choice_c": "[1.09, 1.15]", + "choice_d": "[0.39, 0.54]", + "answer_gt": "[1.09, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 55 + }, + { + "path": "3528-168669-0060.flac", + "question": "Listen for the occurrence of 'am' in the recording.", + "choice_a": "[0.84, 0.92]", + "choice_b": "[0.43, 0.58]", + "choice_c": "[1.39, 1.56]", + "choice_d": "[0.58, 0.68]", + "answer_gt": "[0.58, 0.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 56 + }, + { + "path": "1998-29454-0040.flac", + "question": "What is the duration of 'think' being mentioned in the conversation?", + "choice_a": "[1.29, 1.46]", + "choice_b": "[0.94, 1.29]", + "choice_c": "[1.92, 2.44]", + "choice_d": "[1.29, 1.46]", + "answer_gt": "[0.94, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 57 + }, + { + "path": "6841-88291-0004.flac", + "question": "Determine the precise moment 'the' in the audio's pace.", + "choice_a": "[0.81, 1.02]", + "choice_b": "[0.37, 0.63]", + "choice_c": "[1.02, 1.05]", + "choice_d": "[0.30, 0.37]", + "answer_gt": "[0.30, 0.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 58 + }, + { + "path": "3764-168671-0052.flac", + "question": "Identify the time segment when 'drink' is mentioned in the audio.", + "choice_a": "[2.86, 2.94]", + "choice_b": "[5.81, 6.08]", + "choice_c": "[2.97, 3.42]", + "choice_d": "[1.67, 2.20]", + "answer_gt": "[2.97, 3.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 59 + }, + { + "path": "7018-75789-0023.flac", + "question": "Recognize the segment where 'hundred' is spoken by the speaker.", + "choice_a": "[1.82, 1.93]", + "choice_b": "[1.54, 1.82]", + "choice_c": "[1.11, 1.17]", + "choice_d": "[0.81, 1.11]", + "answer_gt": "[1.54, 1.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 60 + }, + { + "path": "2609-157645-0001.flac", + "question": "Locate the segment where 'be' appears most frequently.", + "choice_a": "[0.46, 1.28]", + "choice_b": "[1.34, 1.77]", + "choice_c": "[6.27, 6.43]", + "choice_d": "[0.32, 0.46]", + "answer_gt": "[6.27, 6.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 61 + }, + { + "path": "2277-149896-0010.flac", + "question": "What is the timestamp for when 'little' occurs?", + "choice_a": "[1.06, 1.50]", + "choice_b": "[0.48, 0.81]", + "choice_c": "[0.81, 0.99]", + "choice_d": "[3.67, 3.94]", + "answer_gt": "[0.48, 0.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 62 + }, + { + "path": "6070-86744-0002.flac", + "question": "Recognize the segment where 'madrid' is spoken by the speaker.", + "choice_a": "[8.75, 9.29]", + "choice_b": "[1.06, 11.4]", + "choice_c": "[7.61, 7.75]", + "choice_d": "[4.86, 5.57]", + "answer_gt": "[8.75, 9.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 63 + }, + { + "path": "6241-66616-0016.flac", + "question": "At what time interval does 'going' occur in the audio clip?", + "choice_a": "[1.12, 1.23]", + "choice_b": "[3.79, 4.06]", + "choice_c": "[3.46, 3.64]", + "choice_d": "[2.58, 2.63]", + "answer_gt": "[3.79, 4.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 64 + }, + { + "path": "8224-274381-0000.flac", + "question": "Determine the time when 'discouraged' happens within the clip.", + "choice_a": "[5.44, 16.1]", + "choice_b": "[7.13, 8.01]", + "choice_c": "[1.87, 2.04]", + "choice_d": "[0.32, 10.5]", + "answer_gt": "[7.13, 8.01]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 65 + }, + { + "path": "1686-142278-0031.flac", + "question": "Note the time when 'dear' is first mentioned in the audio.", + "choice_a": "[0.92, 1.12]", + "choice_b": "[0.54, 0.92]", + "choice_c": "[2.41, 2.67]", + "choice_d": "[3.38, 3.62]", + "answer_gt": "[0.92, 1.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 66 + }, + { + "path": "6313-66125-0008.flac", + "question": "Identify the moment 'guide' is mentioned in the audio.", + "choice_a": "[1.41, 1.87]", + "choice_b": "[0.97, 1.13]", + "choice_c": "[3.42, 3.75]", + "choice_d": "[1.93, 2.33]", + "answer_gt": "[1.93, 2.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 67 + }, + { + "path": "6930-76324-0001.flac", + "question": "Determine the point in the audio when 'their' is discussed.", + "choice_a": "[1.74, 2.29]", + "choice_b": "[1.32, 1.61]", + "choice_c": "[2.41, 2.51]", + "choice_d": "[0.31, 0.38]", + "answer_gt": "[2.41, 2.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 68 + }, + { + "path": "260-123286-0000.flac", + "question": "Listen for the occurrence of 'saturday' in the recording.", + "choice_a": "[4.71, 5.06]", + "choice_b": "[0.56, 0.99]", + "choice_c": "[6.22, 6.31]", + "choice_d": "[3.43, 3.49]", + "answer_gt": "[0.56, 0.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 69 + }, + { + "path": "5694-64029-0032.flac", + "question": "What is the timestamp for when 'reported' occurs?", + "choice_a": "[1.54, 1.99]", + "choice_b": "[2.05, 2.33]", + "choice_c": "[0.91, 1.03]", + "choice_d": "[1.99, 2.05]", + "answer_gt": "[1.54, 1.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 70 + }, + { + "path": "2414-128292-0007.flac", + "question": "Locate the segment where 'doth' appears most frequently.", + "choice_a": "[0.83, 1.09]", + "choice_b": "[3.11, 3.27]", + "choice_c": "[2.49, 3.11]", + "choice_d": "[3.27, 3.34]", + "answer_gt": "[0.83, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 71 + }, + { + "path": "8230-279154-0037.flac", + "question": "In the audio track, when is 'an' first noticed?", + "choice_a": "[7.57, 17.7]", + "choice_b": "[2.61, 12.7]", + "choice_c": "[7.73, 18.5]", + "choice_d": "[5.91, 16.8]", + "answer_gt": "[2.61, 12.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 72 + }, + { + "path": "5484-24318-0014.flac", + "question": "Establish when 'him' is first mentioned in the clip.", + "choice_a": "[2.55, 2.68]", + "choice_b": "[5.21, 5.65]", + "choice_c": "[3.54, 3.86]", + "choice_d": "[2.83, 3.02]", + "answer_gt": "[3.54, 3.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 73 + }, + { + "path": "4153-186223-0018.flac", + "question": "What is the timestamp for the initial occurrence of 'said'?", + "choice_a": "[0.77, 1.07]", + "choice_b": "[1.60, 1.66]", + "choice_c": "[1.39, 1.60]", + "choice_d": "[0.58, 0.77]", + "answer_gt": "[1.39, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 74 + }, + { + "path": "4852-28319-0019.flac", + "question": "Highlight the time when 'and' is mentioned.", + "choice_a": "[4.97, 5.40]", + "choice_b": "[0.54, 0.64]", + "choice_c": "[9.84, 10.1]", + "choice_d": "[3.65, 3.76]", + "answer_gt": "[3.65, 3.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 75 + }, + { + "path": "7641-96252-0020.flac", + "question": "Determine the precise moment 'that' in the audio's pace.", + "choice_a": "[3.14, 3.34]", + "choice_b": "[0.24, 0.46]", + "choice_c": "[1.81, 2.05]", + "choice_d": "[2.35, 2.80]", + "answer_gt": "[3.14, 3.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 76 + }, + { + "path": "4570-24733-0000.flac", + "question": "Pinpoint the passage where 'words' is heard in the background.", + "choice_a": "[6.76, 7.16]", + "choice_b": "[1.58, 12.1]", + "choice_c": "[3.87, 4.18]", + "choice_d": "[8.04, 8.78]", + "answer_gt": "[1.58, 12.1]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 77 + }, + { + "path": "2609-157645-0006.flac", + "question": "Highlight the time when 'a' is mentioned.", + "choice_a": "[0.88, 1.03]", + "choice_b": "[1.31, 1.36]", + "choice_c": "[1.31, 1.36]", + "choice_d": "[8.04, 8.33]", + "answer_gt": "[1.31, 1.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 78 + }, + { + "path": "6345-93306-0015.flac", + "question": "When does the audio feature the sound of 'was'?", + "choice_a": "[7.31, 7.48]", + "choice_b": "[3.72, 3.81]", + "choice_c": "[1.77, 1.97]", + "choice_d": "[5.59, 5.71]", + "answer_gt": "[7.31, 7.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 79 + }, + { + "path": "1673-143396-0007.flac", + "question": "Listen for the precise interval when 'pravity' is dominant.", + "choice_a": "[3.20, 3.71]", + "choice_b": "[0.41, 0.57]", + "choice_c": "[1.67, 1.79]", + "choice_d": "[4.58, 5.13]", + "answer_gt": "[3.20, 3.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 80 + }, + { + "path": "3536-23268-0005.flac", + "question": "Highlight the time when 'that' is mentioned.", + "choice_a": "[1.83, 1.96]", + "choice_b": "[1.83, 1.96]", + "choice_c": "[2.82, 2.97]", + "choice_d": "[1.14, 1.57]", + "answer_gt": "[2.82, 2.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 81 + }, + { + "path": "2086-149220-0041.flac", + "question": "Find out when the speaker refers to 'of' during the audio.", + "choice_a": "[3.49, 3.61]", + "choice_b": "[3.49, 3.61]", + "choice_c": "[0.53, 0.66]", + "choice_d": "[2.72, 2.78]", + "answer_gt": "[3.49, 3.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 82 + }, + { + "path": "2961-961-0010.flac", + "question": "At what point in the audio is 'by' clearly audible?", + "choice_a": "[2.65, 2.81]", + "choice_b": "[1.81, 1.87]", + "choice_c": "[4.53, 4.72]", + "choice_d": "[0.61, 0.75]", + "answer_gt": "[4.53, 4.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 83 + }, + { + "path": "1650-167613-0029.flac", + "question": "Find out the time when 'truth' is immediately occured.", + "choice_a": "[0.61, 0.77]", + "choice_b": "[0.77, 1.13]", + "choice_c": "[0.61, 0.77]", + "choice_d": "[0.77, 1.13]", + "answer_gt": "[0.77, 1.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 84 + }, + { + "path": "6267-53049-0023.flac", + "question": "Listen for when 'gave' is emphasized by the narrator.", + "choice_a": "[3.51, 4.01]", + "choice_b": "[0.59, 0.90]", + "choice_c": "[0.96, 1.57]", + "choice_d": "[1.72, 2.06]", + "answer_gt": "[1.72, 2.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 85 + }, + { + "path": "260-123288-0025.flac", + "question": "In the audio track, when is 'with' first noticed?", + "choice_a": "[0.65, 0.80]", + "choice_b": "[8.86, 8.99]", + "choice_c": "[3.15, 3.41]", + "choice_d": "[2.95, 3.12]", + "answer_gt": "[8.86, 8.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 86 + }, + { + "path": "3764-168671-0030.flac", + "question": "When is 'little' first heard in the recording?", + "choice_a": "[1.11, 1.39]", + "choice_b": "[1.01, 1.11]", + "choice_c": "[2.38, 2.64]", + "choice_d": "[3.27, 3.54]", + "answer_gt": "[1.11, 1.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 87 + }, + { + "path": "6467-94831-0002.flac", + "question": "Locate the segment where 'mister' appears most frequently.", + "choice_a": "[0.52, 0.77]", + "choice_b": "[1.38, 1.63]", + "choice_c": "[0.27, 0.52]", + "choice_d": "[0.83, 0.98]", + "answer_gt": "[1.38, 1.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 88 + }, + { + "path": "6345-93302-0022.flac", + "question": "Pinpoint the exact moment 'only' appears.", + "choice_a": "[1.04, 1.27]", + "choice_b": "[0.90, 1.04]", + "choice_c": "[1.78, 2.02]", + "choice_d": "[1.27, 1.66]", + "answer_gt": "[1.04, 1.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 89 + }, + { + "path": "1650-173551-0007.flac", + "question": "Recognize the segment where 'bazdeev's' is spoken by the speaker.", + "choice_a": "[2.97, 3.66]", + "choice_b": "[0.76, 1.00]", + "choice_c": "[1.11, 1.67]", + "choice_d": "[2.78, 2.97]", + "answer_gt": "[2.97, 3.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 90 + }, + { + "path": "8463-294828-0015.flac", + "question": "At what interval does 'time' appear in the audio?", + "choice_a": "[1.87, 2.20]", + "choice_b": "[2.20, 2.62]", + "choice_c": "[1.83, 1.87]", + "choice_d": "[0.95, 1.37]", + "answer_gt": "[2.20, 2.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 91 + }, + { + "path": "1686-142278-0037.flac", + "question": "Determine the point in the audio when 'with' is discussed.", + "choice_a": "[2.10, 2.19]", + "choice_b": "[4.53, 4.64]", + "choice_c": "[1.80, 2.10]", + "choice_d": "[1.74, 1.80]", + "answer_gt": "[4.53, 4.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 92 + }, + { + "path": "8254-84205-0036.flac", + "question": "Establish when 'be' is first mentioned in the clip.", + "choice_a": "[2.13, 2.25]", + "choice_b": "[0.78, 1.10]", + "choice_c": "[1.63, 1.66]", + "choice_d": "[2.69, 2.98]", + "answer_gt": "[2.13, 2.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 93 + }, + { + "path": "4153-186222-0016.flac", + "question": "Locate the segment where 'say' appears most frequently.", + "choice_a": "[0.41, 0.60]", + "choice_b": "[2.20, 2.37]", + "choice_c": "[0.28, 0.41]", + "choice_d": "[1.43, 1.92]", + "answer_gt": "[0.41, 0.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 94 + }, + { + "path": "908-157963-0000.flac", + "question": "Find the moment 'heard' is clearly audible in the audio.", + "choice_a": "[6.73, 7.10]", + "choice_b": "[8.72, 9.00]", + "choice_c": "[7.38, 7.81]", + "choice_d": "[1.62, 2.05]", + "answer_gt": "[7.38, 7.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 95 + }, + { + "path": "8463-294828-0005.flac", + "question": "Identify the moment 'my' is mentioned in the audio.", + "choice_a": "[1.18, 1.40]", + "choice_b": "[0.25, 0.90]", + "choice_c": "[0.25, 0.90]", + "choice_d": "[0.25, 0.90]", + "answer_gt": "[1.18, 1.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 96 + }, + { + "path": "1995-1836-0004_0.flac", + "question": "Identify the moment 'his' is mentioned in the audio.", + "choice_a": "[2.86, 3.09]", + "choice_b": "[7.85, 18.0]", + "choice_c": "[7.30, 7.70]", + "choice_d": "[6.18, 26.6]", + "answer_gt": "[7.85, 18.0]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 97 + }, + { + "path": "652-130726-0010.flac", + "question": "Determine the point in the audio when 'for' is discussed.", + "choice_a": "[3.05, 3.22]", + "choice_b": "[1.82, 1.95]", + "choice_c": "[1.53, 1.82]", + "choice_d": "[4.42, 4.93]", + "answer_gt": "[1.82, 1.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 98 + }, + { + "path": "7902-96594-0010.flac", + "question": "Pinpoint the exact moment 'didn't' appears.", + "choice_a": "[0.75, 1.06]", + "choice_b": "[1.78, 1.93]", + "choice_c": "[1.06, 1.39]", + "choice_d": "[1.43, 1.64]", + "answer_gt": "[1.43, 1.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 99 + }, + { + "path": "4350-9170-0058.flac", + "question": "Listen for when 'and' is emphasized by the narrator.", + "choice_a": "[0.22, 0.33]", + "choice_b": "[3.31, 3.48]", + "choice_c": "[1.89, 2.01]", + "choice_d": "[0.33, 0.70]", + "answer_gt": "[1.89, 2.01]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 100 + }, + { + "path": "4198-12259-0002.flac", + "question": "Pinpoint the passage where 'with' is heard in the background.", + "choice_a": "[1.35, 1.59]", + "choice_b": "[1.59, 2.03]", + "choice_c": "[2.03, 2.27]", + "choice_d": "[0.37, 0.49]", + "answer_gt": "[2.03, 2.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 101 + }, + { + "path": "3752-4943-0008.flac", + "question": "When does the speaker mention 'white'?", + "choice_a": "[0.81, 0.86]", + "choice_b": "[0.19, 0.42]", + "choice_c": "[0.86, 1.43]", + "choice_d": "[1.59, 2.03]", + "answer_gt": "[1.59, 2.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 102 + }, + { + "path": "1688-142285-0071.flac", + "question": "Establish when 'put' is first mentioned in the clip.", + "choice_a": "[0.53, 0.75]", + "choice_b": "[2.14, 2.49]", + "choice_c": "[2.49, 2.98]", + "choice_d": "[0.75, 0.94]", + "answer_gt": "[0.53, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 103 + }, + { + "path": "4153-185072-0009.flac", + "question": "Recognize the segment where 'found' is spoken by the speaker.", + "choice_a": "[5.41, 5.65]", + "choice_b": "[4.42, 4.58]", + "choice_c": "[1.52, 1.69]", + "choice_d": "[7.03, 7.38]", + "answer_gt": "[7.03, 7.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 104 + }, + { + "path": "6432-63723-0053.flac", + "question": "Listen to this audio clip; identify the time segment when 'disentangle' is spoken.", + "choice_a": "[5.71, 5.83]", + "choice_b": "[0.73, 0.88]", + "choice_c": "[3.79, 4.61]", + "choice_d": "[4.66, 4.99]", + "answer_gt": "[3.79, 4.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 105 + }, + { + "path": "6432-63722-0028.flac", + "question": "Note the time interval that includes the mention of 'yes'.", + "choice_a": "[2.56, 2.76]", + "choice_b": "[2.26, 2.56]", + "choice_c": "[1.47, 1.70]", + "choice_d": "[0.24, 0.40]", + "answer_gt": "[2.26, 2.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 106 + }, + { + "path": "7697-105815-0019.flac", + "question": "Highlight when 'our' is spoken.", + "choice_a": "[0.71, 1.10]", + "choice_b": "[1.10, 1.22]", + "choice_c": "[1.22, 1.40]", + "choice_d": "[0.47, 0.71]", + "answer_gt": "[0.47, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 107 + }, + { + "path": "5639-40744-0040.flac", + "question": "Identify the time segment when 'sufficient' is mentioned in the audio.", + "choice_a": "[5.90, 6.58]", + "choice_b": "[5.10, 5.66]", + "choice_c": "[4.28, 4.36]", + "choice_d": "[2.63, 2.79]", + "answer_gt": "[5.10, 5.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 108 + }, + { + "path": "4515-11057-0103.flac", + "question": "Establish when 'can't' is first mentioned in the clip.", + "choice_a": "[1.11, 1.43]", + "choice_b": "[0.77, 1.11]", + "choice_c": "[3.26, 3.75]", + "choice_d": "[1.85, 2.17]", + "answer_gt": "[3.26, 3.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 109 + }, + { + "path": "2428-83705-0007.flac", + "question": "In the audio track, when is 'don't' first noticed?", + "choice_a": "[0.45, 0.75]", + "choice_b": "[1.97, 2.17]", + "choice_c": "[0.87, 1.41]", + "choice_d": "[0.75, 0.87]", + "answer_gt": "[1.97, 2.17]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 110 + }, + { + "path": "251-118436-0018.flac", + "question": "At what time interval does 'know' occur in the audio clip?", + "choice_a": "[0.55, 0.82]", + "choice_b": "[0.43, 0.55]", + "choice_c": "[1.11, 1.38]", + "choice_d": "[1.71, 1.82]", + "answer_gt": "[0.55, 0.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 111 + }, + { + "path": "6432-63722-0010.flac", + "question": "Highlight when 'we' is spoken.", + "choice_a": "[1.96, 2.08]", + "choice_b": "[0.54, 0.69]", + "choice_c": "[1.41, 1.52]", + "choice_d": "[0.69, 0.92]", + "answer_gt": "[1.96, 2.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 112 + }, + { + "path": "237-134493-0018.flac", + "question": "When is 'under' first heard in the recording?", + "choice_a": "[1.77, 1.86]", + "choice_b": "[3.33, 3.56]", + "choice_c": "[0.38, 0.59]", + "choice_d": "[0.77, 1.09]", + "answer_gt": "[3.33, 3.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 113 + }, + { + "path": "1188-133604-0028.flac", + "question": "Ascertain the point at which 'it' enters the discussion.", + "choice_a": "[6.43, 6.49]", + "choice_b": "[9.27, 9.96]", + "choice_c": "[4.22, 14.3]", + "choice_d": "[8.70, 8.84]", + "answer_gt": "[4.22, 14.3]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 114 + }, + { + "path": "7176-92135-0014.flac", + "question": "Find out the time when 'opposition' is immediately occured.", + "choice_a": "[9.80, 10.4]", + "choice_b": "[2.98, 3.42]", + "choice_c": "[1.95, 2.07]", + "choice_d": "[8.29, 8.71]", + "answer_gt": "[9.80, 10.4]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 115 + }, + { + "path": "8288-274162-0028.flac", + "question": "Locate the segment where 'only' appears most frequently.", + "choice_a": "[0.58, 0.82]", + "choice_b": "[1.53, 1.80]", + "choice_c": "[1.29, 1.42]", + "choice_d": "[1.42, 1.53]", + "answer_gt": "[1.53, 1.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 116 + }, + { + "path": "84-121550-0025.flac", + "question": "Highlight when 'awhile' is spoken.", + "choice_a": "[3.97, 4.23]", + "choice_b": "[4.23, 4.54]", + "choice_c": "[5.85, 6.45]", + "choice_d": "[3.66, 3.97]", + "answer_gt": "[5.85, 6.45]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 117 + }, + { + "path": "2035-147961-0012.flac", + "question": "When in the audio does 'for' coincide with a specific event?", + "choice_a": "[2.91, 3.08]", + "choice_b": "[1.19, 1.58]", + "choice_c": "[2.47, 2.91]", + "choice_d": "[1.58, 1.72]", + "answer_gt": "[1.58, 1.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 118 + }, + { + "path": "4198-12281-0015.flac", + "question": "At what point in the audio is 'john' clearly audible?", + "choice_a": "[0.58, 10.8]", + "choice_b": "[0.88, 11.4]", + "choice_c": "[2.09, 2.61]", + "choice_d": "[7.08, 7.20]", + "answer_gt": "[2.09, 2.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 119 + }, + { + "path": "3080-5040-0016.flac", + "question": "Recognize the time when 'it' is mentioned by the speaker.", + "choice_a": "[0.61, 0.78]", + "choice_b": "[1.87, 1.95]", + "choice_c": "[0.42, 0.49]", + "choice_d": "[1.05, 1.18]", + "answer_gt": "[1.05, 1.18]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 120 + }, + { + "path": "7975-280085-0017.flac", + "question": "Recognize the time when 'including' is mentioned by the speaker.", + "choice_a": "[1.84, 2.03]", + "choice_b": "[2.03, 2.21]", + "choice_c": "[0.23, 0.67]", + "choice_d": "[2.21, 2.81]", + "answer_gt": "[0.23, 0.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 121 + }, + { + "path": "8280-266249-0064.flac", + "question": "Highlight the time when 'in' is mentioned.", + "choice_a": "[1.22, 1.64]", + "choice_b": "[1.02, 1.22]", + "choice_c": "[0.78, 1.02]", + "choice_d": "[0.24, 0.43]", + "answer_gt": "[1.02, 1.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 122 + }, + { + "path": "652-130726-0029.flac", + "question": "Find out when the speaker refers to 'the' during the audio.", + "choice_a": "[0.64, 0.69]", + "choice_b": "[3.98, 4.16]", + "choice_c": "[6.61, 6.87]", + "choice_d": "[0.64, 0.69]", + "answer_gt": "[0.64, 0.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 123 + }, + { + "path": "7975-280085-0018.flac", + "question": "Ascertain the exact time 'order' is mentioned in the audio.", + "choice_a": "[0.25, 0.39]", + "choice_b": "[0.69, 1.12]", + "choice_c": "[1.12, 1.37]", + "choice_d": "[5.20, 5.46]", + "answer_gt": "[1.12, 1.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 124 + }, + { + "path": "8463-294825-0002.flac", + "question": "Ascertain the exact time 'the' is mentioned in the audio.", + "choice_a": "[0.38, 0.81]", + "choice_b": "[0.81, 0.95]", + "choice_c": "[9.30, 9.38]", + "choice_d": "[4.55, 5.22]", + "answer_gt": "[9.30, 9.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 125 + }, + { + "path": "1701-141760-0012.flac", + "question": "Highlight when 'half' is spoken.", + "choice_a": "[1.60, 1.92]", + "choice_b": "[1.21, 1.33]", + "choice_c": "[1.92, 1.97]", + "choice_d": "[6.66, 6.94]", + "answer_gt": "[1.60, 1.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 126 + }, + { + "path": "251-136532-0001.flac", + "question": "Find out the time when 'had' is immediately occured.", + "choice_a": "[1.45, 2.19]", + "choice_b": "[1.39, 1.45]", + "choice_c": "[2.57, 2.68]", + "choice_d": "[3.74, 3.97]", + "answer_gt": "[2.57, 2.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 127 + }, + { + "path": "4294-9934-0012.flac", + "question": "When in the audio does 'said' coincide with a specific event?", + "choice_a": "[1.93, 2.17]", + "choice_b": "[0.87, 0.98]", + "choice_c": "[0.98, 1.32]", + "choice_d": "[1.46, 1.73]", + "answer_gt": "[1.93, 2.17]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 128 + }, + { + "path": "2035-147961-0019.flac", + "question": "Identify the portion of the clip where 'howls' is most prominent.", + "choice_a": "[1.48, 1.71]", + "choice_b": "[1.85, 2.34]", + "choice_c": "[1.15, 1.48]", + "choice_d": "[0.81, 1.07]", + "answer_gt": "[0.81, 1.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 129 + }, + { + "path": "6455-67803-0025.flac", + "question": "Establish when 'lady' is first mentioned in the clip.", + "choice_a": "[2.38, 2.60]", + "choice_b": "[0.94, 1.30]", + "choice_c": "[1.55, 1.66]", + "choice_d": "[1.55, 1.66]", + "answer_gt": "[2.38, 2.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 130 + }, + { + "path": "3752-4944-0019.flac", + "question": "Identify the portion of the clip where 'at' is most prominent.", + "choice_a": "[1.78, 1.89]", + "choice_b": "[0.40, 0.51]", + "choice_c": "[0.75, 1.09]", + "choice_d": "[1.22, 1.78]", + "answer_gt": "[1.78, 1.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 131 + }, + { + "path": "3528-168669-0028.flac", + "question": "Listen to this audio clip: when does 'then' occur?", + "choice_a": "[0.71, 1.11]", + "choice_b": "[2.10, 2.22]", + "choice_c": "[1.70, 2.01]", + "choice_d": "[2.22, 2.62]", + "answer_gt": "[0.71, 1.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 132 + }, + { + "path": "84-121550-0031.flac", + "question": "Pinpoint the exact moment 'as' is repeated in the audio.", + "choice_a": "[2.75, 3.11]", + "choice_b": "[4.54, 4.95]", + "choice_c": "[3.96, 4.12]", + "choice_d": "[0.39, 0.54]", + "answer_gt": "[0.39, 0.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 133 + }, + { + "path": "2277-149874-0009.flac", + "question": "When does 'wrapped' occur in the audio?", + "choice_a": "[0.38, 0.54]", + "choice_b": "[1.02, 1.21]", + "choice_c": "[0.93, 1.02]", + "choice_d": "[1.66, 1.95]", + "answer_gt": "[1.66, 1.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 134 + }, + { + "path": "2961-960-0006.flac", + "question": "Pinpoint the exact moment 'prevailed' appears.", + "choice_a": "[2.84, 13.0]", + "choice_b": "[8.22, 9.02]", + "choice_c": "[9.75, 10.2]", + "choice_d": "[0.55, 1.15]", + "answer_gt": "[8.22, 9.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 135 + }, + { + "path": "7601-175351-0017.flac", + "question": "Pinpoint the exact moment 'ought' is repeated in the audio.", + "choice_a": "[0.94, 1.18]", + "choice_b": "[0.83, 0.94]", + "choice_c": "[1.18, 1.62]", + "choice_d": "[0.49, 0.83]", + "answer_gt": "[0.49, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 136 + }, + { + "path": "4515-11057-0040.flac", + "question": "Pinpoint the exact moment 'well' is repeated in the audio.", + "choice_a": "[1.83, 1.96]", + "choice_b": "[0.34, 0.77]", + "choice_c": "[1.33, 1.49]", + "choice_d": "[1.74, 1.83]", + "answer_gt": "[0.34, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 137 + }, + { + "path": "1089-134691-0013.flac", + "question": "Determine the point in the audio when 'him' is discussed.", + "choice_a": "[9.71, 9.82]", + "choice_b": "[5.91, 5.98]", + "choice_c": "[7.14, 7.31]", + "choice_d": "[3.13, 13.2]", + "answer_gt": "[3.13, 13.2]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 138 + }, + { + "path": "2086-149220-0021.flac", + "question": "In the audio track, when is 'eyes' first noticed?", + "choice_a": "[1.27, 1.47]", + "choice_b": "[1.80, 2.13]", + "choice_c": "[2.43, 2.71]", + "choice_d": "[0.39, 0.47]", + "answer_gt": "[1.80, 2.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 139 + }, + { + "path": "1630-96099-0010.flac", + "question": "Locate the part of the audio where 'this' is expressed.", + "choice_a": "[1.25, 1.43]", + "choice_b": "[1.93, 2.09]", + "choice_c": "[3.81, 4.03]", + "choice_d": "[0.23, 0.32]", + "answer_gt": "[3.81, 4.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 140 + }, + { + "path": "6267-53049-0010.flac", + "question": "Ascertain the exact time 'once' is mentioned in the audio.", + "choice_a": "[3.31, 3.54]", + "choice_b": "[2.04, 2.62]", + "choice_c": "[2.79, 3.08]", + "choice_d": "[0.18, 0.33]", + "answer_gt": "[2.04, 2.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 141 + }, + { + "path": "8131-117016-0050.flac", + "question": "Ascertain the point at which 'a' enters the discussion.", + "choice_a": "[0.55, 0.75]", + "choice_b": "[0.96, 1.03]", + "choice_c": "[0.75, 0.96]", + "choice_d": "[0.75, 0.96]", + "answer_gt": "[0.96, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 142 + }, + { + "path": "6841-88291-0025.flac", + "question": "Determine the point in the audio when 'he' is discussed.", + "choice_a": "[0.71, 0.86]", + "choice_b": "[0.38, 0.42]", + "choice_c": "[0.91, 1.33]", + "choice_d": "[0.28, 0.38]", + "answer_gt": "[0.38, 0.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 143 + }, + { + "path": "2035-152373-0009.flac", + "question": "Identify the portion of the clip where 'apostles' is most prominent.", + "choice_a": "[3.46, 3.62]", + "choice_b": "[5.18, 5.61]", + "choice_c": "[2.92, 13.4]", + "choice_d": "[5.61, 5.69]", + "answer_gt": "[2.92, 13.4]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 144 + }, + { + "path": "1673-143396-0000.flac", + "question": "Find out the time when 'the' is immediately occured.", + "choice_a": "[1.48, 1.62]", + "choice_b": "[1.62, 1.75]", + "choice_c": "[6.28, 6.79]", + "choice_d": "[1.12, 11.3]", + "answer_gt": "[1.62, 1.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 145 + }, + { + "path": "8463-287645-0010.flac", + "question": "Ascertain the exact time 'he' is mentioned in the audio.", + "choice_a": "[0.26, 0.38]", + "choice_b": "[0.38, 0.71]", + "choice_c": "[2.04, 2.37]", + "choice_d": "[1.17, 1.57]", + "answer_gt": "[0.26, 0.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 146 + }, + { + "path": "4852-28330-0010.flac", + "question": "Pinpoint the exact moment 'moment' appears.", + "choice_a": "[2.14, 2.46]", + "choice_b": "[2.79, 2.83]", + "choice_c": "[2.83, 3.19]", + "choice_d": "[3.19, 3.53]", + "answer_gt": "[3.19, 3.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 147 + }, + { + "path": "4970-29095-0021.flac", + "question": "Recognize the time when 'signing' is most clearly articulated.", + "choice_a": "[5.00, 5.48]", + "choice_b": "[0.35, 10.8]", + "choice_c": "[8.82, 8.95]", + "choice_d": "[2.50, 2.72]", + "answer_gt": "[0.35, 10.8]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 148 + }, + { + "path": "1089-134686-0015.flac", + "question": "Ascertain the point at which 'dusk' enters the discussion.", + "choice_a": "[0.12, 0.26]", + "choice_b": "[0.32, 0.82]", + "choice_c": "[3.45, 3.97]", + "choice_d": "[0.26, 0.32]", + "answer_gt": "[0.32, 0.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 149 + }, + { + "path": "8455-210777-0029.flac", + "question": "Determine the point in the audio when 'to' is repeated.", + "choice_a": "[2.05, 2.30]", + "choice_b": "[0.90, 1.00]", + "choice_c": "[1.00, 1.26]", + "choice_d": "[2.30, 2.40]", + "answer_gt": "[2.30, 2.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 150 + }, + { + "path": "2803-161169-0012.flac", + "question": "When is 'this' first mentioned in the clip?", + "choice_a": "[8.23, 8.49]", + "choice_b": "[7.48, 7.67]", + "choice_c": "[6.35, 6.71]", + "choice_d": "[0.46, 0.78]", + "answer_gt": "[8.23, 8.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 151 + }, + { + "path": "700-122866-0012.flac", + "question": "Find out the time when 'to' is immediately occured.", + "choice_a": "[5.03, 5.13]", + "choice_b": "[4.10, 4.26]", + "choice_c": "[0.62, 1.02]", + "choice_d": "[2.55, 2.77]", + "answer_gt": "[5.03, 5.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 152 + }, + { + "path": "2506-13150-0006.flac", + "question": "Note the time interval that includes the mention of 'meet'.", + "choice_a": "[2.06, 2.39]", + "choice_b": "[2.73, 2.92]", + "choice_c": "[2.49, 2.73]", + "choice_d": "[0.32, 0.46]", + "answer_gt": "[2.06, 2.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 153 + }, + { + "path": "2428-83699-0011.flac", + "question": "When in the audio does 'it' coincide with a specific event?", + "choice_a": "[1.17, 1.33]", + "choice_b": "[1.33, 1.80]", + "choice_c": "[0.30, 0.49]", + "choice_d": "[0.91, 1.03]", + "answer_gt": "[0.91, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 154 + }, + { + "path": "6467-97061-0022.flac", + "question": "Identify the portion of the clip where 'they' is most prominent.", + "choice_a": "[6.06, 6.28]", + "choice_b": "[5.90, 6.06]", + "choice_c": "[6.94, 7.19]", + "choice_d": "[1.74, 1.82]", + "answer_gt": "[1.74, 1.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 155 + }, + { + "path": "1255-90407-0029.flac", + "question": "Determine the point in the audio when 'gone' is discussed.", + "choice_a": "[3.43, 3.62]", + "choice_b": "[0.60, 0.67]", + "choice_c": "[1.70, 1.83]", + "choice_d": "[3.31, 3.43]", + "answer_gt": "[3.43, 3.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 156 + }, + { + "path": "2094-142345-0027.flac", + "question": "Determine the precise moment 'munny' in the audio's pace.", + "choice_a": "[2.39, 2.66]", + "choice_b": "[2.66, 2.74]", + "choice_c": "[0.53, 1.16]", + "choice_d": "[1.39, 1.59]", + "answer_gt": "[0.53, 1.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 157 + }, + { + "path": "4077-13754-0002.flac", + "question": "Highlight when 'interests' is spoken.", + "choice_a": "[2.72, 3.26]", + "choice_b": "[9.20, 9.70]", + "choice_c": "[2.65, 2.72]", + "choice_d": "[5.75, 16.1]", + "answer_gt": "[5.75, 16.1]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 158 + }, + { + "path": "4198-61336-0006.flac", + "question": "Find out when the speaker refers to 'of' during the audio.", + "choice_a": "[5.41, 5.58]", + "choice_b": "[4.81, 5.41]", + "choice_c": "[3.54, 3.82]", + "choice_d": "[1.68, 2.03]", + "answer_gt": "[5.41, 5.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 159 + }, + { + "path": "7127-75947-0032.flac", + "question": "Locate the segment where 'what' appears most frequently.", + "choice_a": "[3.47, 3.63]", + "choice_b": "[1.86, 1.93]", + "choice_c": "[3.63, 3.82]", + "choice_d": "[0.38, 0.77]", + "answer_gt": "[3.63, 3.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 160 + }, + { + "path": "6345-93302-0003.flac", + "question": "Find out when the speaker refers to 'environment' during the audio.", + "choice_a": "[1.05, 1.40]", + "choice_b": "[1.40, 1.77]", + "choice_c": "[3.16, 3.88]", + "choice_d": "[0.96, 1.05]", + "answer_gt": "[3.16, 3.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 161 + }, + { + "path": "237-126133-0002.flac", + "question": "What is the duration of 'said' being mentioned in the conversation?", + "choice_a": "[4.71, 4.84]", + "choice_b": "[1.07, 1.33]", + "choice_c": "[5.26, 5.43]", + "choice_d": "[3.45, 3.59]", + "answer_gt": "[1.07, 1.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 162 + }, + { + "path": "8173-294714-0058.flac", + "question": "In the audio track, when is 'my' first noticed?", + "choice_a": "[1.72, 2.41]", + "choice_b": "[0.21, 0.32]", + "choice_c": "[1.60, 1.72]", + "choice_d": "[1.47, 1.60]", + "answer_gt": "[1.60, 1.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 163 + }, + { + "path": "3660-172183-0017.flac", + "question": "Pinpoint the exact moment 'through' is repeated in the audio.", + "choice_a": "[5.83, 5.93]", + "choice_b": "[2.65, 2.79]", + "choice_c": "[0.41, 0.52]", + "choice_d": "[5.47, 5.83]", + "answer_gt": "[2.65, 2.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 164 + }, + { + "path": "8455-210777-0034.flac", + "question": "Note the time when 'it' is first mentioned in the audio.", + "choice_a": "[2.99, 3.26]", + "choice_b": "[5.61, 5.82]", + "choice_c": "[3.26, 3.84]", + "choice_d": "[5.82, 5.90]", + "answer_gt": "[5.82, 5.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 165 + }, + { + "path": "6829-68771-0005.flac", + "question": "What is the timestamp for the initial occurrence of 'and'?", + "choice_a": "[3.63, 3.80]", + "choice_b": "[0.64, 0.90]", + "choice_c": "[1.81, 2.04]", + "choice_d": "[4.31, 4.71]", + "answer_gt": "[3.63, 3.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 166 + }, + { + "path": "121-121726-0009.flac", + "question": "Identify the portion of the clip where 'up' is most prominent.", + "choice_a": "[2.26, 2.50]", + "choice_b": "[4.04, 4.34]", + "choice_c": "[4.34, 4.54]", + "choice_d": "[4.90, 5.43]", + "answer_gt": "[4.34, 4.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 167 + }, + { + "path": "8254-84205-0018.flac", + "question": "Identify the portion of the clip where 'been' is most prominent.", + "choice_a": "[0.25, 0.49]", + "choice_b": "[1.92, 2.12]", + "choice_c": "[2.81, 2.89]", + "choice_d": "[2.12, 2.36]", + "answer_gt": "[1.92, 2.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 168 + }, + { + "path": "1272-141231-0022.flac", + "question": "Note the time interval that includes the mention of 'an'.", + "choice_a": "[1.24, 1.55]", + "choice_b": "[4.90, 5.09]", + "choice_c": "[3.09, 3.31]", + "choice_d": "[1.55, 1.69]", + "answer_gt": "[1.55, 1.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 169 + }, + { + "path": "3528-168669-0014.flac", + "question": "Find the duration when 'in' occurs consecutively in the audio.", + "choice_a": "[1.02, 1.11]", + "choice_b": "[0.80, 0.85]", + "choice_c": "[0.48, 0.68]", + "choice_d": "[0.85, 1.02]", + "answer_gt": "[1.02, 1.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 170 + }, + { + "path": "1580-141084-0042.flac", + "question": "Find out when the speaker refers to 'to' during the audio.", + "choice_a": "[2.68, 3.15]", + "choice_b": "[1.97, 2.07]", + "choice_c": "[3.67, 4.16]", + "choice_d": "[1.79, 1.97]", + "answer_gt": "[1.97, 2.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 171 + }, + { + "path": "1462-170145-0015.flac", + "question": "Highlight the time when 'bridge' is mentioned.", + "choice_a": "[1.61, 2.13]", + "choice_b": "[0.53, 0.67]", + "choice_c": "[3.54, 3.64]", + "choice_d": "[0.32, 0.50]", + "answer_gt": "[1.61, 2.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 172 + }, + { + "path": "3853-163249-0025.flac", + "question": "When in the audio does 'quiet' coincide with a specific event?", + "choice_a": "[2.28, 2.67]", + "choice_b": "[0.78, 1.27]", + "choice_c": "[4.71, 4.94]", + "choice_d": "[3.40, 3.50]", + "answer_gt": "[2.28, 2.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 173 + }, + { + "path": "1919-142785-0020.flac", + "question": "Pinpoint the exact moment 'a' appears.", + "choice_a": "[5.18, 5.22]", + "choice_b": "[6.81, 7.21]", + "choice_c": "[4.24, 4.80]", + "choice_d": "[3.34, 3.43]", + "answer_gt": "[5.18, 5.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 174 + }, + { + "path": "5543-27761-0053.flac", + "question": "Listen for when 'fedosya' is emphasized by the narrator.", + "choice_a": "[1.90, 2.38]", + "choice_b": "[1.47, 1.73]", + "choice_c": "[0.55, 0.63]", + "choice_d": "[3.02, 3.53]", + "answer_gt": "[3.02, 3.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 175 + }, + { + "path": "7641-96684-0015.flac", + "question": "Identify the time segment when 'that' is mentioned in the audio.", + "choice_a": "[5.71, 5.93]", + "choice_b": "[1.26, 1.53]", + "choice_c": "[5.31, 5.43]", + "choice_d": "[1.19, 1.26]", + "answer_gt": "[5.31, 5.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 176 + }, + { + "path": "3853-163249-0033.flac", + "question": "Listen for when 'ann' is emphasized by the narrator.", + "choice_a": "[0.68, 0.98]", + "choice_b": "[1.47, 1.52]", + "choice_c": "[0.46, 0.68]", + "choice_d": "[2.56, 2.91]", + "answer_gt": "[2.56, 2.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 177 + }, + { + "path": "1701-141760-0040.flac", + "question": "Establish when 'do' is first mentioned in the clip.", + "choice_a": "[3.08, 3.57]", + "choice_b": "[4.51, 5.22]", + "choice_c": "[0.60, 10.8]", + "choice_d": "[3.57, 3.71]", + "answer_gt": "[0.60, 10.8]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 178 + }, + { + "path": "700-122866-0016.flac", + "question": "Listen for the occurrence of 'oh' in the recording.", + "choice_a": "[1.91, 2.06]", + "choice_b": "[1.34, 1.74]", + "choice_c": "[2.06, 2.14]", + "choice_d": "[0.44, 0.70]", + "answer_gt": "[0.44, 0.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 179 + }, + { + "path": "3853-163249-0029.flac", + "question": "When is 'vittles' first mentioned in the clip?", + "choice_a": "[0.78, 1.15]", + "choice_b": "[1.64, 2.00]", + "choice_c": "[2.90, 3.41]", + "choice_d": "[2.24, 2.41]", + "answer_gt": "[1.64, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 180 + }, + { + "path": "1580-141084-0030.flac", + "question": "Note the time interval that includes the mention of 'said'.", + "choice_a": "[1.24, 1.48]", + "choice_b": "[0.58, 0.83]", + "choice_c": "[0.83, 0.92]", + "choice_d": "[0.36, 0.58]", + "answer_gt": "[1.24, 1.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 181 + }, + { + "path": "3528-168656-0007.flac", + "question": "Pinpoint the passage where 'subject' is heard in the background.", + "choice_a": "[2.95, 3.17]", + "choice_b": "[1.33, 1.70]", + "choice_c": "[4.53, 4.58]", + "choice_d": "[2.85, 2.95]", + "answer_gt": "[1.33, 1.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 182 + }, + { + "path": "6467-94831-0040.flac", + "question": "Highlight the time when 'i' is mentioned.", + "choice_a": "[2.93, 3.06]", + "choice_b": "[2.20, 2.36]", + "choice_c": "[0.64, 1.23]", + "choice_d": "[1.26, 1.44]", + "answer_gt": "[2.20, 2.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 183 + }, + { + "path": "1701-141760-0043.flac", + "question": "Note the time interval that includes the mention of 'in'.", + "choice_a": "[5.28, 5.38]", + "choice_b": "[4.12, 4.64]", + "choice_c": "[3.64, 4.12]", + "choice_d": "[5.93, 5.97]", + "answer_gt": "[5.28, 5.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 184 + }, + { + "path": "2428-83699-0007.flac", + "question": "Listen for the precise interval when 'to' is dominant.", + "choice_a": "[4.88, 4.98]", + "choice_b": "[2.50, 2.69]", + "choice_c": "[0.94, 1.63]", + "choice_d": "[0.15, 0.25]", + "answer_gt": "[4.88, 4.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 185 + }, + { + "path": "1255-90407-0005.flac", + "question": "Listen for when 'should' is emphasized by the narrator.", + "choice_a": "[0.55, 0.64]", + "choice_b": "[0.81, 1.05]", + "choice_c": "[0.64, 0.81]", + "choice_d": "[0.81, 1.05]", + "answer_gt": "[0.64, 0.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 186 + }, + { + "path": "4831-18525-0027.flac", + "question": "At what point in the audio is 'as' clearly audible?", + "choice_a": "[0.35, 0.52]", + "choice_b": "[1.89, 2.07]", + "choice_c": "[0.97, 1.45]", + "choice_d": "[2.51, 2.92]", + "answer_gt": "[1.89, 2.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 187 + }, + { + "path": "2609-156975-0007.flac", + "question": "Pinpoint the exact moment 'of' is repeated in the audio.", + "choice_a": "[5.02, 5.16]", + "choice_b": "[4.32, 14.7]", + "choice_c": "[5.02, 5.16]", + "choice_d": "[5.21, 15.3]", + "answer_gt": "[5.02, 5.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 188 + }, + { + "path": "116-288048-0001.flac", + "question": "Find the duration when 'born' occurs consecutively in the audio.", + "choice_a": "[1.74, 2.05]", + "choice_b": "[0.58, 0.70]", + "choice_c": "[0.44, 0.58]", + "choice_d": "[1.35, 1.51]", + "answer_gt": "[1.74, 2.05]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 189 + }, + { + "path": "4446-2275-0031.flac", + "question": "Note the time when 'i' is first mentioned in the audio.", + "choice_a": "[0.55, 0.74]", + "choice_b": "[2.26, 2.48]", + "choice_c": "[1.26, 1.72]", + "choice_d": "[0.55, 0.74]", + "answer_gt": "[0.55, 0.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 190 + }, + { + "path": "3997-180294-0026.flac", + "question": "Identify the portion of the clip where 'the' is most prominent.", + "choice_a": "[7.15, 7.65]", + "choice_b": "[1.03, 1.17]", + "choice_c": "[5.42, 5.98]", + "choice_d": "[0.46, 0.53]", + "answer_gt": "[0.46, 0.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 191 + }, + { + "path": "1701-141760-0001.flac", + "question": "Spot the exact moment 'their' appears in the audio.", + "choice_a": "[1.46, 1.70]", + "choice_b": "[0.99, 1.08]", + "choice_c": "[3.16, 3.50]", + "choice_d": "[1.32, 1.46]", + "answer_gt": "[1.32, 1.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 192 + }, + { + "path": "4323-18416-0020.flac", + "question": "Find out the time when 'if' is immediately occured.", + "choice_a": "[7.54, 7.80]", + "choice_b": "[0.53, 0.71]", + "choice_c": "[6.27, 6.34]", + "choice_d": "[1.97, 2.05]", + "answer_gt": "[0.53, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 193 + }, + { + "path": "1988-147956-0019.flac", + "question": "Discover when 'of' is introduced in the conversation.", + "choice_a": "[1.56, 1.82]", + "choice_b": "[4.32, 4.76]", + "choice_c": "[1.82, 1.89]", + "choice_d": "[4.87, 5.05]", + "answer_gt": "[1.82, 1.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 194 + }, + { + "path": "4852-28330-0020.flac", + "question": "Pinpoint the exact moment 'during' is repeated in the audio.", + "choice_a": "[3.43, 3.49]", + "choice_b": "[0.60, 0.69]", + "choice_c": "[1.01, 1.43]", + "choice_d": "[0.49, 0.60]", + "answer_gt": "[1.01, 1.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 195 + }, + { + "path": "8461-281231-0002.flac", + "question": "Determine the point in the audio when 'with' is repeated.", + "choice_a": "[2.39, 2.55]", + "choice_b": "[6.08, 6.59]", + "choice_c": "[4.42, 4.86]", + "choice_d": "[2.04, 2.39]", + "answer_gt": "[2.39, 2.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 196 + }, + { + "path": "2414-128292-0016.flac", + "question": "Ascertain the point at which 'perhaps' enters the discussion.", + "choice_a": "[1.60, 2.07]", + "choice_b": "[1.48, 1.60]", + "choice_c": "[0.88, 1.44]", + "choice_d": "[0.33, 0.55]", + "answer_gt": "[1.60, 2.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 197 + }, + { + "path": "2830-3980-0051.flac", + "question": "Detect the point at which 'justification' is spoken.", + "choice_a": "[3.23, 4.13]", + "choice_b": "[1.19, 1.44]", + "choice_c": "[1.44, 1.49]", + "choice_d": "[1.85, 1.97]", + "answer_gt": "[3.23, 4.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 198 + }, + { + "path": "4831-25894-0009.flac", + "question": "When is 'her' first mentioned in the clip?", + "choice_a": "[1.29, 11.5]", + "choice_b": "[2.52, 2.80]", + "choice_c": "[6.11, 6.46]", + "choice_d": "[0.71, 0.78]", + "answer_gt": "[0.71, 0.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 199 + }, + { + "path": "6295-244435-0032.flac", + "question": "Listen to this audio clip; identify the time segment when 'it' is spoken.", + "choice_a": "[0.12, 0.27]", + "choice_b": "[2.30, 2.39]", + "choice_c": "[2.39, 2.89]", + "choice_d": "[2.13, 2.30]", + "answer_gt": "[0.12, 0.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 200 + }, + { + "path": "2830-3980-0037.flac", + "question": "At what time interval does 'the' occur in the audio clip?", + "choice_a": "[3.01, 3.38]", + "choice_b": "[0.95, 1.04]", + "choice_c": "[0.81, 0.95]", + "choice_d": "[0.95, 1.04]", + "answer_gt": "[0.95, 1.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 201 + }, + { + "path": "4153-186223-0008.flac", + "question": "Pinpoint the exact moment 'your' is repeated in the audio.", + "choice_a": "[2.10, 2.34]", + "choice_b": "[2.03, 2.10]", + "choice_c": "[0.60, 0.82]", + "choice_d": "[0.82, 1.27]", + "answer_gt": "[2.03, 2.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 202 + }, + { + "path": "6123-59150-0029.flac", + "question": "Find the duration when 'faces' occurs consecutively in the audio.", + "choice_a": "[1.81, 1.89]", + "choice_b": "[0.49, 0.75]", + "choice_c": "[2.60, 2.72]", + "choice_d": "[1.36, 1.81]", + "answer_gt": "[1.36, 1.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 203 + }, + { + "path": "174-50561-0011.flac", + "question": "When is 'dresses' first mentioned in the clip?", + "choice_a": "[7.97, 8.33]", + "choice_b": "[8.51, 8.69]", + "choice_c": "[0.57, 0.65]", + "choice_d": "[1.72, 2.36]", + "answer_gt": "[1.72, 2.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 204 + }, + { + "path": "7850-111771-0005.flac", + "question": "Highlight when 'federal' is spoken.", + "choice_a": "[1.72, 2.00]", + "choice_b": "[0.47, 0.60]", + "choice_c": "[0.60, 0.77]", + "choice_d": "[2.11, 2.43]", + "answer_gt": "[2.11, 2.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 205 + }, + { + "path": "7176-92135-0042.flac", + "question": "Pinpoint the exact moment 'do' appears.", + "choice_a": "[0.87, 1.22]", + "choice_b": "[5.12, 5.27]", + "choice_c": "[0.19, 0.38]", + "choice_d": "[4.04, 4.17]", + "answer_gt": "[5.12, 5.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 206 + }, + { + "path": "3660-6517-0002.flac", + "question": "Recognize the segment where 'quarter' is spoken by the speaker.", + "choice_a": "[4.96, 5.43]", + "choice_b": "[3.96, 4.14]", + "choice_c": "[1.74, 1.89]", + "choice_d": "[2.00, 2.30]", + "answer_gt": "[4.96, 5.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 207 + }, + { + "path": "4572-112383-0001.flac", + "question": "Pinpoint the passage where 'where' is heard in the background.", + "choice_a": "[0.52, 0.97]", + "choice_b": "[0.52, 0.97]", + "choice_c": "[1.26, 1.44]", + "choice_d": "[1.26, 1.44]", + "answer_gt": "[0.52, 0.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 208 + }, + { + "path": "237-134500-0002.flac", + "question": "Determine the point in the audio when 'wind' is discussed.", + "choice_a": "[0.31, 0.40]", + "choice_b": "[3.71, 4.05]", + "choice_c": "[2.11, 2.57]", + "choice_d": "[0.77, 1.05]", + "answer_gt": "[0.77, 1.05]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 209 + }, + { + "path": "1993-147965-0005.flac", + "question": "Ascertain the exact time 'about' is mentioned in the audio.", + "choice_a": "[7.48, 7.71]", + "choice_b": "[3.51, 4.14]", + "choice_c": "[4.32, 4.61]", + "choice_d": "[0.65, 0.96]", + "answer_gt": "[0.65, 0.96]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 210 + }, + { + "path": "4446-2275-0030.flac", + "question": "At what time interval does 'said' occur in the audio clip?", + "choice_a": "[0.62, 0.83]", + "choice_b": "[1.58, 1.84]", + "choice_c": "[0.27, 0.62]", + "choice_d": "[0.83, 1.00]", + "answer_gt": "[1.58, 1.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 211 + }, + { + "path": "8288-274162-0051.flac", + "question": "In the audio track, when is 'then' first noticed?", + "choice_a": "[1.11, 1.35]", + "choice_b": "[0.72, 1.11]", + "choice_c": "[0.66, 0.72]", + "choice_d": "[0.66, 0.72]", + "answer_gt": "[1.11, 1.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 212 + }, + { + "path": "1284-1180-0000.flac", + "question": "Identify the moment 'of' is mentioned in the audio.", + "choice_a": "[0.36, 0.59]", + "choice_b": "[6.15, 6.32]", + "choice_c": "[0.85, 1.14]", + "choice_d": "[3.01, 3.16]", + "answer_gt": "[6.15, 6.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 213 + }, + { + "path": "700-122866-0023.flac", + "question": "Listen to this audio clip: when does 'did' occur?", + "choice_a": "[0.32, 0.41]", + "choice_b": "[0.32, 0.41]", + "choice_c": "[0.70, 0.98]", + "choice_d": "[1.56, 1.88]", + "answer_gt": "[1.56, 1.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 214 + }, + { + "path": "6467-97061-0006.flac", + "question": "At what time interval does 'terrible' occur in the audio clip?", + "choice_a": "[2.26, 2.40]", + "choice_b": "[0.45, 0.84]", + "choice_c": "[3.40, 3.95]", + "choice_d": "[4.95, 5.13]", + "answer_gt": "[0.45, 0.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 215 + }, + { + "path": "3576-138058-0001.flac", + "question": "At what time interval does 'don' occur in the audio clip?", + "choice_a": "[8.23, 8.66]", + "choice_b": "[0.38, 0.52]", + "choice_c": "[1.96, 2.12]", + "choice_d": "[0.52, 0.71]", + "answer_gt": "[0.52, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 216 + }, + { + "path": "3005-163390-0013.flac", + "question": "Find the duration when 'well' occurs consecutively in the audio.", + "choice_a": "[1.78, 1.94]", + "choice_b": "[0.88, 1.23]", + "choice_c": "[0.52, 0.73]", + "choice_d": "[3.46, 3.63]", + "answer_gt": "[0.52, 0.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 217 + }, + { + "path": "6313-66125-0016.flac", + "question": "Determine the precise moment 'the' in the audio's pace.", + "choice_a": "[5.07, 5.37]", + "choice_b": "[1.58, 1.64]", + "choice_c": "[3.49, 3.78]", + "choice_d": "[1.58, 1.64]", + "answer_gt": "[1.58, 1.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 218 + }, + { + "path": "1320-122617-0039.flac", + "question": "Ascertain the point at which 'mohican' enters the discussion.", + "choice_a": "[1.57, 1.94]", + "choice_b": "[2.54, 2.90]", + "choice_c": "[5.65, 5.88]", + "choice_d": "[0.34, 0.83]", + "answer_gt": "[0.34, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 219 + }, + { + "path": "3000-15664-0003.flac", + "question": "Find the moment 'go' is clearly audible in the audio.", + "choice_a": "[0.47, 0.63]", + "choice_b": "[0.63, 1.35]", + "choice_c": "[3.15, 3.25]", + "choice_d": "[3.67, 4.06]", + "answer_gt": "[0.47, 0.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 220 + }, + { + "path": "3764-168670-0019.flac", + "question": "Pinpoint the exact moment 'what' is repeated in the audio.", + "choice_a": "[0.90, 1.12]", + "choice_b": "[1.12, 1.33]", + "choice_c": "[0.53, 0.90]", + "choice_d": "[1.36, 1.67]", + "answer_gt": "[0.53, 0.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 221 + }, + { + "path": "4323-55228-0028.flac", + "question": "Find out the time when 'billy' is immediately occured.", + "choice_a": "[1.72, 1.84]", + "choice_b": "[0.80, 1.08]", + "choice_c": "[0.76, 0.80]", + "choice_d": "[1.11, 1.35]", + "answer_gt": "[1.11, 1.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 222 + }, + { + "path": "2414-128292-0019.flac", + "question": "When in the audio does 'how' coincide with a specific event?", + "choice_a": "[1.17, 1.69]", + "choice_b": "[0.51, 0.71]", + "choice_c": "[0.71, 1.03]", + "choice_d": "[0.71, 1.03]", + "answer_gt": "[0.51, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 223 + }, + { + "path": "2830-3980-0013.flac", + "question": "Note the time when 'because' is first mentioned in the audio.", + "choice_a": "[2.23, 2.35]", + "choice_b": "[1.90, 2.16]", + "choice_c": "[0.46, 0.58]", + "choice_d": "[2.35, 2.76]", + "answer_gt": "[1.90, 2.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 224 + }, + { + "path": "5543-27761-0069.flac", + "question": "Highlight when 'it' is spoken.", + "choice_a": "[1.31, 1.61]", + "choice_b": "[1.78, 2.45]", + "choice_c": "[0.79, 0.85]", + "choice_d": "[0.56, 0.63]", + "answer_gt": "[0.56, 0.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 225 + }, + { + "path": "6467-94831-0004.flac", + "question": "When is 'i' first heard in the recording?", + "choice_a": "[1.56, 1.74]", + "choice_b": "[0.86, 1.35]", + "choice_c": "[4.47, 4.93]", + "choice_d": "[3.13, 3.50]", + "answer_gt": "[1.56, 1.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 226 + }, + { + "path": "5442-32873-0012.flac", + "question": "Recognize the time when 'here' is mentioned by the speaker.", + "choice_a": "[4.30, 4.42]", + "choice_b": "[0.92, 1.27]", + "choice_c": "[2.10, 2.16]", + "choice_d": "[2.53, 2.82]", + "answer_gt": "[0.92, 1.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 227 + }, + { + "path": "4446-2273-0033.flac", + "question": "Pinpoint the passage where 'time' is heard in the background.", + "choice_a": "[0.96, 1.34]", + "choice_b": "[1.54, 1.87]", + "choice_c": "[0.77, 0.96]", + "choice_d": "[2.07, 2.44]", + "answer_gt": "[0.96, 1.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 228 + }, + { + "path": "6128-63244-0015.flac", + "question": "Determine the point in the audio when 'a' is discussed.", + "choice_a": "[4.14, 4.27]", + "choice_b": "[3.13, 3.43]", + "choice_c": "[0.28, 0.47]", + "choice_d": "[1.92, 1.98]", + "answer_gt": "[1.92, 1.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 229 + }, + { + "path": "1993-147149-0004.flac", + "question": "Listen for when 'days' is emphasized by the narrator.", + "choice_a": "[2.08, 2.35]", + "choice_b": "[3.24, 3.79]", + "choice_c": "[5.53, 5.83]", + "choice_d": "[2.64, 2.79]", + "answer_gt": "[2.08, 2.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 230 + }, + { + "path": "7176-88083-0011.flac", + "question": "In the audio track, when is 'it' first noticed?", + "choice_a": "[3.06, 3.16]", + "choice_b": "[7.45, 8.00]", + "choice_c": "[2.04, 2.17]", + "choice_d": "[2.62, 2.69]", + "answer_gt": "[2.04, 2.17]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 231 + }, + { + "path": "7641-96670-0006.flac", + "question": "At what point in the audio is 'of' clearly audible?", + "choice_a": "[3.28, 3.71]", + "choice_b": "[0.35, 0.71]", + "choice_c": "[3.10, 3.22]", + "choice_d": "[2.28, 2.64]", + "answer_gt": "[3.10, 3.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 232 + }, + { + "path": "3528-168669-0121.flac", + "question": "Listen to this audio clip: when does 'will' occur?", + "choice_a": "[1.38, 1.54]", + "choice_b": "[2.02, 2.09]", + "choice_c": "[1.29, 1.38]", + "choice_d": "[0.47, 0.76]", + "answer_gt": "[1.29, 1.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 233 + }, + { + "path": "8297-275154-0013.flac", + "question": "Find the duration when 'nor' occurs consecutively in the audio.", + "choice_a": "[1.36, 1.48]", + "choice_b": "[2.11, 2.30]", + "choice_c": "[0.54, 0.64]", + "choice_d": "[0.64, 0.80]", + "answer_gt": "[2.11, 2.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 234 + }, + { + "path": "2609-169640-0020.flac", + "question": "What is the duration of 'john' being mentioned in the conversation?", + "choice_a": "[1.72, 2.14]", + "choice_b": "[0.51, 0.89]", + "choice_c": "[1.64, 1.72]", + "choice_d": "[1.00, 1.64]", + "answer_gt": "[1.72, 2.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 235 + }, + { + "path": "5683-32865-0004.flac", + "question": "Highlight the time when 'with' is mentioned.", + "choice_a": "[5.54, 5.69]", + "choice_b": "[3.52, 3.66]", + "choice_c": "[2.65, 3.10]", + "choice_d": "[3.86, 4.66]", + "answer_gt": "[5.54, 5.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 236 + }, + { + "path": "6319-64726-0005.flac", + "question": "Identify when 'pierced' is heard in the audio.", + "choice_a": "[6.19, 6.39]", + "choice_b": "[6.39, 6.76]", + "choice_c": "[5.65, 6.19]", + "choice_d": "[0.55, 0.76]", + "answer_gt": "[5.65, 6.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 237 + }, + { + "path": "533-131562-0000.flac", + "question": "Recognize the segment where 'said' is spoken by the speaker.", + "choice_a": "[2.86, 3.28]", + "choice_b": "[0.68, 1.01]", + "choice_c": "[4.98, 15.0]", + "choice_d": "[3.60, 4.05]", + "answer_gt": "[2.86, 3.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 238 + }, + { + "path": "2277-149897-0033.flac", + "question": "Identify the time segment when 'another' is mentioned in the audio.", + "choice_a": "[1.22, 1.50]", + "choice_b": "[4.19, 4.35]", + "choice_c": "[3.18, 3.34]", + "choice_d": "[2.36, 2.50]", + "answer_gt": "[1.22, 1.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 239 + }, + { + "path": "2414-159411-0034.flac", + "question": "Determine the time when 'journey' happens within the clip.", + "choice_a": "[1.15, 1.25]", + "choice_b": "[1.25, 1.79]", + "choice_c": "[1.00, 1.15]", + "choice_d": "[1.89, 2.17]", + "answer_gt": "[1.25, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 240 + }, + { + "path": "1221-135766-0003.flac", + "question": "Listen for the occurrence of 'if' in the recording.", + "choice_a": "[8.70, 9.23]", + "choice_b": "[0.06, 10.1]", + "choice_c": "[3.53, 3.78]", + "choice_d": "[7.15, 7.48]", + "answer_gt": "[0.06, 10.1]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 241 + }, + { + "path": "3576-138058-0022.flac", + "question": "What is the duration of 'i' being mentioned in the conversation?", + "choice_a": "[9.98, 20.1]", + "choice_b": "[6.51, 16.6]", + "choice_c": "[7.50, 7.53]", + "choice_d": "[5.27, 5.74]", + "answer_gt": "[7.50, 7.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 242 + }, + { + "path": "2300-131720-0025.flac", + "question": "When is 'they' first heard in the recording?", + "choice_a": "[3.69, 14.2]", + "choice_b": "[1.02, 11.3]", + "choice_c": "[0.83, 1.10]", + "choice_d": "[0.54, 0.67]", + "answer_gt": "[0.54, 0.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 243 + }, + { + "path": "5543-27761-0017.flac", + "question": "Pinpoint the exact moment 'of' appears.", + "choice_a": "[6.40, 7.08]", + "choice_b": "[9.54, 9.77]", + "choice_c": "[4.37, 4.50]", + "choice_d": "[2.01, 2.09]", + "answer_gt": "[4.37, 4.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 244 + }, + { + "path": "2609-169640-0013.flac", + "question": "Recognize the time when 'course' is mentioned by the speaker.", + "choice_a": "[1.89, 2.44]", + "choice_b": "[1.75, 1.89]", + "choice_c": "[1.15, 1.39]", + "choice_d": "[1.42, 1.75]", + "answer_gt": "[1.89, 2.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 245 + }, + { + "path": "6345-93306-0007.flac", + "question": "When is 'is' first mentioned in the clip?", + "choice_a": "[3.92, 4.30]", + "choice_b": "[4.33, 4.57]", + "choice_c": "[1.46, 1.57]", + "choice_d": "[1.46, 1.57]", + "answer_gt": "[1.46, 1.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 246 + }, + { + "path": "3660-172183-0001.flac", + "question": "When does the speaker mention 'steeds'?", + "choice_a": "[6.29, 6.75]", + "choice_b": "[7.02, 7.45]", + "choice_c": "[3.49, 3.97]", + "choice_d": "[3.17, 3.24]", + "answer_gt": "[3.49, 3.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 247 + }, + { + "path": "2035-147961-0016.flac", + "question": "Listen for the precise interval when 'his' is dominant.", + "choice_a": "[1.22, 1.48]", + "choice_b": "[1.70, 1.88]", + "choice_c": "[2.32, 2.46]", + "choice_d": "[0.76, 0.82]", + "answer_gt": "[2.32, 2.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 248 + }, + { + "path": "700-122867-0003.flac", + "question": "Listen to this audio clip; identify the time segment when 'you're' is spoken.", + "choice_a": "[0.96, 1.09]", + "choice_b": "[1.54, 1.64]", + "choice_c": "[1.09, 1.54]", + "choice_d": "[1.64, 1.94]", + "answer_gt": "[0.96, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 249 + }, + { + "path": "1580-141084-0039.flac", + "question": "At what interval does 'and' appear in the audio?", + "choice_a": "[1.99, 2.13]", + "choice_b": "[0.53, 1.04]", + "choice_c": "[2.98, 3.08]", + "choice_d": "[1.15, 1.29]", + "answer_gt": "[1.15, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 250 + }, + { + "path": "8254-84205-0000.flac", + "question": "When is 'nine' first mentioned in the clip?", + "choice_a": "[2.59, 2.72]", + "choice_b": "[0.69, 1.02]", + "choice_c": "[1.02, 1.41]", + "choice_d": "[2.13, 2.55]", + "answer_gt": "[1.02, 1.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 251 + }, + { + "path": "3170-137482-0044.flac", + "question": "Listen to this audio clip: when does 'found' occur?", + "choice_a": "[0.98, 1.26]", + "choice_b": "[2.70, 12.9]", + "choice_c": "[9.09, 9.51]", + "choice_d": "[4.13, 4.27]", + "answer_gt": "[0.98, 1.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 252 + }, + { + "path": "7018-75788-0013.flac", + "question": "What is the timestamp for the initial occurrence of 'to'?", + "choice_a": "[3.72, 3.78]", + "choice_b": "[5.94, 6.36]", + "choice_c": "[8.69, 9.05]", + "choice_d": "[9.57, 9.70]", + "answer_gt": "[9.57, 9.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 253 + }, + { + "path": "8188-269288-0049.flac", + "question": "Find out when the speaker refers to 'can't' during the audio.", + "choice_a": "[0.50, 0.73]", + "choice_b": "[1.10, 1.20]", + "choice_c": "[1.24, 1.79]", + "choice_d": "[2.39, 2.50]", + "answer_gt": "[0.50, 0.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 254 + }, + { + "path": "4077-13751-0009.flac", + "question": "In the audio track, when is 'erie' first noticed?", + "choice_a": "[9.30, 9.64]", + "choice_b": "[3.10, 3.26]", + "choice_c": "[6.54, 6.92]", + "choice_d": "[7.46, 7.90]", + "answer_gt": "[7.46, 7.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 255 + }, + { + "path": "3570-5694-0005.flac", + "question": "Spot the exact moment 'the' appears in the audio.", + "choice_a": "[6.14, 6.35]", + "choice_b": "[5.64, 6.14]", + "choice_c": "[0.68, 0.77]", + "choice_d": "[0.68, 0.77]", + "answer_gt": "[0.68, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 256 + }, + { + "path": "251-137823-0011.flac", + "question": "Recognize the time when 'beam' is mentioned by the speaker.", + "choice_a": "[4.20, 4.59]", + "choice_b": "[3.45, 3.75]", + "choice_c": "[1.96, 2.08]", + "choice_d": "[0.49, 1.01]", + "answer_gt": "[4.20, 4.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 257 + }, + { + "path": "3000-15664-0024.flac", + "question": "Locate the part of the audio where 'crags' is expressed.", + "choice_a": "[8.25, 8.67]", + "choice_b": "[6.04, 6.15]", + "choice_c": "[2.86, 3.29]", + "choice_d": "[1.51, 12.0]", + "answer_gt": "[1.51, 12.0]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 258 + }, + { + "path": "777-126732-0009.flac", + "question": "At what point in the audio is 'head' clearly audible?", + "choice_a": "[2.73, 3.09]", + "choice_b": "[2.34, 2.52]", + "choice_c": "[1.41, 1.71]", + "choice_d": "[2.05, 2.34]", + "answer_gt": "[2.34, 2.52]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 259 + }, + { + "path": "7902-96594-0002.flac", + "question": "In the audio track, when is 'ready' first noticed?", + "choice_a": "[1.08, 1.29]", + "choice_b": "[1.29, 1.64]", + "choice_c": "[0.39, 0.79]", + "choice_d": "[0.39, 0.79]", + "answer_gt": "[1.29, 1.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 260 + }, + { + "path": "251-118436-0009.flac", + "question": "At what time interval does 'in' occur in the audio clip?", + "choice_a": "[0.71, 1.17]", + "choice_b": "[5.41, 5.51]", + "choice_c": "[3.65, 4.18]", + "choice_d": "[0.57, 0.71]", + "answer_gt": "[5.41, 5.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 261 + }, + { + "path": "4350-9170-0059.flac", + "question": "Discover when 'man' is introduced in the conversation.", + "choice_a": "[3.81, 3.88]", + "choice_b": "[0.91, 1.04]", + "choice_c": "[0.81, 0.91]", + "choice_d": "[3.88, 4.23]", + "answer_gt": "[3.88, 4.23]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 262 + }, + { + "path": "8288-274162-0011.flac", + "question": "In the audio track, when is 'i' first noticed?", + "choice_a": "[0.89, 1.00]", + "choice_b": "[1.30, 1.45]", + "choice_c": "[0.53, 0.65]", + "choice_d": "[0.65, 0.85]", + "answer_gt": "[0.53, 0.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 263 + }, + { + "path": "1580-141084-0028.flac", + "question": "Ascertain the exact time 'was' is mentioned in the audio.", + "choice_a": "[1.19, 1.66]", + "choice_b": "[0.60, 0.94]", + "choice_c": "[0.94, 1.19]", + "choice_d": "[0.94, 1.19]", + "answer_gt": "[0.60, 0.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 264 + }, + { + "path": "1998-29454-0036.flac", + "question": "Pinpoint the exact moment 'dickie' is repeated in the audio.", + "choice_a": "[2.80, 3.01]", + "choice_b": "[0.53, 0.79]", + "choice_c": "[0.79, 0.98]", + "choice_d": "[1.03, 1.42]", + "answer_gt": "[1.03, 1.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 265 + }, + { + "path": "3576-138058-0008.flac", + "question": "When is 'him' first mentioned in the clip?", + "choice_a": "[2.12, 2.44]", + "choice_b": "[1.89, 2.06]", + "choice_c": "[0.70, 1.04]", + "choice_d": "[1.55, 1.89]", + "answer_gt": "[1.55, 1.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 266 + }, + { + "path": "1585-131718-0003.flac", + "question": "Locate the segment where 'or' appears most frequently.", + "choice_a": "[5.31, 5.83]", + "choice_b": "[0.01, 10.5]", + "choice_c": "[5.83, 5.95]", + "choice_d": "[4.15, 4.27]", + "answer_gt": "[5.83, 5.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 267 + }, + { + "path": "3752-4944-0029.flac", + "question": "Pinpoint the exact moment 'whom' appears.", + "choice_a": "[1.20, 1.56]", + "choice_b": "[1.14, 1.20]", + "choice_c": "[0.52, 0.77]", + "choice_d": "[0.77, 0.88]", + "answer_gt": "[0.52, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 268 + }, + { + "path": "4992-41797-0009.flac", + "question": "Identify when 'but' is heard in the audio.", + "choice_a": "[4.64, 5.23]", + "choice_b": "[8.11, 8.36]", + "choice_c": "[2.92, 3.14]", + "choice_d": "[0.80, 1.25]", + "answer_gt": "[8.11, 8.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 269 + }, + { + "path": "7729-102255-0031.flac", + "question": "Locate the segment where 'pretended' appears most frequently.", + "choice_a": "[1.79, 2.35]", + "choice_b": "[1.03, 1.21]", + "choice_c": "[2.65, 2.76]", + "choice_d": "[1.21, 1.79]", + "answer_gt": "[1.21, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 270 + }, + { + "path": "1995-1826-0024.flac", + "question": "Pinpoint the passage where 'silver' is heard in the background.", + "choice_a": "[2.06, 2.26]", + "choice_b": "[0.85, 1.57]", + "choice_c": "[4.15, 4.31]", + "choice_d": "[2.39, 2.83]", + "answer_gt": "[2.39, 2.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 271 + }, + { + "path": "7601-175351-0019.flac", + "question": "Identify the time segment when 'every' is mentioned in the audio.", + "choice_a": "[2.89, 3.02]", + "choice_b": "[6.52, 6.76]", + "choice_c": "[0.48, 0.61]", + "choice_d": "[5.42, 15.7]", + "answer_gt": "[5.42, 15.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 272 + }, + { + "path": "8297-275155-0001.flac", + "question": "What is the timestamp for when 'weekly' occurs?", + "choice_a": "[1.80, 2.16]", + "choice_b": "[0.78, 1.13]", + "choice_c": "[5.25, 5.53]", + "choice_d": "[1.29, 1.80]", + "answer_gt": "[1.80, 2.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 273 + }, + { + "path": "5536-43358-0003.flac", + "question": "When does the audio feature the sound of 'might'?", + "choice_a": "[2.72, 2.93]", + "choice_b": "[3.49, 3.68]", + "choice_c": "[0.92, 1.15]", + "choice_d": "[1.15, 1.64]", + "answer_gt": "[0.92, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 274 + }, + { + "path": "7975-280076-0022.flac", + "question": "At what point in the audio is 'was' clearly audible?", + "choice_a": "[0.80, 1.22]", + "choice_b": "[0.63, 0.80]", + "choice_c": "[0.34, 0.63]", + "choice_d": "[1.22, 1.38]", + "answer_gt": "[1.22, 1.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 275 + }, + { + "path": "6829-68769-0018.flac", + "question": "Note the time when 'the' is first mentioned in the audio.", + "choice_a": "[2.56, 2.62]", + "choice_b": "[4.39, 4.77]", + "choice_c": "[4.87, 5.01]", + "choice_d": "[3.23, 3.79]", + "answer_gt": "[2.56, 2.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 276 + }, + { + "path": "4294-14317-0014_0.flac", + "question": "Locate the segment where 'for' appears most frequently.", + "choice_a": "[4.77, 4.92]", + "choice_b": "[2.19, 2.32]", + "choice_c": "[4.71, 24.8]", + "choice_d": "[2.57, 22.7]", + "answer_gt": "[4.77, 4.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 277 + }, + { + "path": "1651-136854-0007.flac", + "question": "Locate the part of the audio where 'publick' is expressed.", + "choice_a": "[7.25, 8.04]", + "choice_b": "[2.06, 2.39]", + "choice_c": "[0.17, 0.47]", + "choice_d": "[0.56, 11.1]", + "answer_gt": "[0.17, 0.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 278 + }, + { + "path": "5895-34622-0022.flac", + "question": "Find out the time when 'itinerant' is immediately occured.", + "choice_a": "[1.53, 1.71]", + "choice_b": "[5.86, 6.48]", + "choice_c": "[7.10, 7.69]", + "choice_d": "[3.61, 3.99]", + "answer_gt": "[5.86, 6.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 279 + }, + { + "path": "4077-13754-0013.flac", + "question": "When does 'this' occur in the audio?", + "choice_a": "[1.34, 11.6]", + "choice_b": "[0.95, 1.57]", + "choice_c": "[0.46, 0.77]", + "choice_d": "[0.77, 0.95]", + "answer_gt": "[0.77, 0.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 280 + }, + { + "path": "1580-141084-0019.flac", + "question": "Identify when 'have' is heard in the audio.", + "choice_a": "[1.72, 1.86]", + "choice_b": "[1.86, 2.24]", + "choice_c": "[0.85, 1.01]", + "choice_d": "[1.01, 1.20]", + "answer_gt": "[1.72, 1.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 281 + }, + { + "path": "4323-18416-0034.flac", + "question": "Identify when 'i've' is heard in the audio.", + "choice_a": "[1.39, 1.53]", + "choice_b": "[3.52, 3.61]", + "choice_c": "[0.37, 0.53]", + "choice_d": "[0.20, 0.37]", + "answer_gt": "[0.20, 0.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 282 + }, + { + "path": "5442-41168-0019.flac", + "question": "Find out the time when 'a' is immediately occured.", + "choice_a": "[2.88, 3.28]", + "choice_b": "[1.08, 1.12]", + "choice_c": "[1.94, 2.51]", + "choice_d": "[1.12, 1.62]", + "answer_gt": "[1.08, 1.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 283 + }, + { + "path": "4515-11057-0101.flac", + "question": "Determine the point in the audio when 'then' is discussed.", + "choice_a": "[0.37, 0.76]", + "choice_b": "[1.25, 1.77]", + "choice_c": "[2.27, 2.41]", + "choice_d": "[2.41, 2.76]", + "answer_gt": "[2.27, 2.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 284 + }, + { + "path": "8288-274162-0022.flac", + "question": "Find out when the speaker refers to 'says' during the audio.", + "choice_a": "[1.09, 1.41]", + "choice_b": "[0.11, 0.29]", + "choice_c": "[2.56, 2.68]", + "choice_d": "[0.78, 1.09]", + "answer_gt": "[0.78, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 285 + }, + { + "path": "8230-279154-0010.flac", + "question": "What is the timestamp for when 'have' occurs?", + "choice_a": "[3.19, 4.02]", + "choice_b": "[8.52, 9.01]", + "choice_c": "[1.18, 1.90]", + "choice_d": "[0.93, 1.10]", + "answer_gt": "[0.93, 1.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 286 + }, + { + "path": "237-134500-0017.flac", + "question": "Pinpoint the exact moment 'if' appears.", + "choice_a": "[0.35, 0.49]", + "choice_b": "[0.35, 0.49]", + "choice_c": "[0.49, 0.74]", + "choice_d": "[0.22, 0.35]", + "answer_gt": "[0.22, 0.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 287 + }, + { + "path": "4198-12259-0008.flac", + "question": "Listen to this audio clip; identify the time segment when 'dead' is spoken.", + "choice_a": "[6.38, 6.68]", + "choice_b": "[4.34, 4.54]", + "choice_c": "[1.45, 12.1]", + "choice_d": "[1.83, 1.94]", + "answer_gt": "[6.38, 6.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 288 + }, + { + "path": "6267-65525-0036.flac", + "question": "Identify the time segment when 'hotel' is mentioned in the audio.", + "choice_a": "[0.41, 0.62]", + "choice_b": "[1.22, 1.57]", + "choice_c": "[1.57, 1.92]", + "choice_d": "[0.80, 1.16]", + "answer_gt": "[0.80, 1.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 289 + }, + { + "path": "672-122797-0002.flac", + "question": "Listen for when 'were' is emphasized by the narrator.", + "choice_a": "[0.59, 0.72]", + "choice_b": "[1.61, 1.84]", + "choice_c": "[0.72, 0.84]", + "choice_d": "[0.49, 10.6]", + "answer_gt": "[0.49, 10.6]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 290 + }, + { + "path": "777-126732-0044.flac", + "question": "Listen to this audio clip; identify the time segment when 'kindred' is spoken.", + "choice_a": "[0.38, 0.56]", + "choice_b": "[1.34, 1.78]", + "choice_c": "[1.18, 1.29]", + "choice_d": "[0.68, 1.13]", + "answer_gt": "[1.34, 1.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 291 + }, + { + "path": "7902-96594-0006.flac", + "question": "Identify when 'shall' is heard in the audio.", + "choice_a": "[0.55, 0.80]", + "choice_b": "[1.08, 1.21]", + "choice_c": "[1.21, 1.46]", + "choice_d": "[0.80, 1.08]", + "answer_gt": "[0.80, 1.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 292 + }, + { + "path": "6123-59150-0016.flac", + "question": "Ascertain the exact time 'fears' is mentioned in the audio.", + "choice_a": "[1.16, 1.54]", + "choice_b": "[1.54, 1.71]", + "choice_c": "[1.78, 2.30]", + "choice_d": "[1.71, 1.78]", + "answer_gt": "[1.78, 2.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 293 + }, + { + "path": "3538-163624-0000.flac", + "question": "Determine the point in the audio when 'king' is discussed.", + "choice_a": "[4.08, 4.20]", + "choice_b": "[2.76, 2.94]", + "choice_c": "[2.01, 2.09]", + "choice_d": "[1.78, 2.01]", + "answer_gt": "[1.78, 2.01]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 294 + }, + { + "path": "5764-299665-0028.flac", + "question": "What is the duration of 'tried' being mentioned in the conversation?", + "choice_a": "[1.23, 1.64]", + "choice_b": "[0.51, 0.99]", + "choice_c": "[3.48, 3.87]", + "choice_d": "[0.99, 1.23]", + "answer_gt": "[0.51, 0.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 295 + }, + { + "path": "3764-168670-0014.flac", + "question": "What is the duration of 'to' being mentioned in the conversation?", + "choice_a": "[4.10, 4.26]", + "choice_b": "[5.95, 6.66]", + "choice_c": "[7.41, 7.76]", + "choice_d": "[1.34, 1.44]", + "answer_gt": "[1.34, 1.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 296 + }, + { + "path": "8455-210777-0024.flac", + "question": "Recognize the segment where 'evil' is spoken by the speaker.", + "choice_a": "[0.83, 0.96]", + "choice_b": "[4.53, 4.63]", + "choice_c": "[0.96, 1.54]", + "choice_d": "[5.40, 5.69]", + "answer_gt": "[0.96, 1.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 297 + }, + { + "path": "3660-172182-0033.flac", + "question": "In the audio track, when is 'the' first noticed?", + "choice_a": "[1.51, 1.58]", + "choice_b": "[2.84, 3.17]", + "choice_c": "[2.05, 2.36]", + "choice_d": "[3.17, 3.22]", + "answer_gt": "[3.17, 3.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 298 + }, + { + "path": "6128-63241-0001.flac", + "question": "Pinpoint the passage where 'the' is heard in the background.", + "choice_a": "[4.39, 5.02]", + "choice_b": "[1.06, 1.30]", + "choice_c": "[3.97, 4.39]", + "choice_d": "[0.53, 0.61]", + "answer_gt": "[0.53, 0.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 299 + }, + { + "path": "7902-96592-0007.flac", + "question": "Discover when 'be' is introduced in the conversation.", + "choice_a": "[2.05, 2.23]", + "choice_b": "[0.94, 1.19]", + "choice_c": "[2.23, 2.41]", + "choice_d": "[1.55, 1.68]", + "answer_gt": "[1.55, 1.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 300 + }, + { + "path": "6930-76324-0008.flac", + "question": "Ascertain the point at which 'of' enters the discussion.", + "choice_a": "[2.25, 2.47]", + "choice_b": "[3.10, 3.52]", + "choice_c": "[3.72, 3.81]", + "choice_d": "[2.85, 3.10]", + "answer_gt": "[3.72, 3.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 301 + }, + { + "path": "3000-15664-0017.flac", + "question": "Find the moment 'sprinkled' is clearly audible in the audio.", + "choice_a": "[6.72, 7.07]", + "choice_b": "[2.68, 13.1]", + "choice_c": "[1.03, 11.7]", + "choice_d": "[9.32, 9.74]", + "answer_gt": "[1.03, 11.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 302 + }, + { + "path": "4570-56594-0017.flac", + "question": "Listen for the precise interval when 'jest' is dominant.", + "choice_a": "[2.46, 2.95]", + "choice_b": "[1.36, 1.43]", + "choice_c": "[0.88, 1.33]", + "choice_d": "[4.77, 5.27]", + "answer_gt": "[2.46, 2.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 303 + }, + { + "path": "422-122949-0027.flac", + "question": "What is the duration of 'with' being mentioned in the conversation?", + "choice_a": "[4.41, 14.5]", + "choice_b": "[1.08, 1.51]", + "choice_c": "[7.50, 7.67]", + "choice_d": "[5.57, 5.68]", + "answer_gt": "[4.41, 14.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 304 + }, + { + "path": "7850-281318-0022.flac", + "question": "Find out when the speaker refers to 'wood' during the audio.", + "choice_a": "[3.14, 3.28]", + "choice_b": "[0.66, 0.98]", + "choice_c": "[1.58, 1.74]", + "choice_d": "[1.74, 1.87]", + "answer_gt": "[0.66, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 305 + }, + { + "path": "7729-102255-0040.flac", + "question": "Identify the portion of the clip where 'triumphant' is most prominent.", + "choice_a": "[3.92, 4.36]", + "choice_b": "[8.85, 9.08]", + "choice_c": "[2.44, 2.52]", + "choice_d": "[0.27, 10.9]", + "answer_gt": "[0.27, 10.9]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 306 + }, + { + "path": "4446-2271-0003.flac", + "question": "Discover when 'it's' is introduced in the conversation.", + "choice_a": "[0.26, 0.41]", + "choice_b": "[0.85, 1.11]", + "choice_c": "[2.40, 2.43]", + "choice_d": "[2.67, 2.99]", + "answer_gt": "[0.26, 0.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 307 + }, + { + "path": "1462-170142-0017.flac", + "question": "Locate the segment where 'get' appears most frequently.", + "choice_a": "[0.57, 0.96]", + "choice_b": "[0.41, 0.57]", + "choice_c": "[0.96, 1.08]", + "choice_d": "[0.33, 0.41]", + "answer_gt": "[0.41, 0.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 308 + }, + { + "path": "8131-117029-0014.flac", + "question": "Spot the exact moment 'and' appears in the audio.", + "choice_a": "[1.40, 1.46]", + "choice_b": "[0.69, 1.10]", + "choice_c": "[2.34, 2.46]", + "choice_d": "[2.46, 2.58]", + "answer_gt": "[2.34, 2.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 309 + }, + { + "path": "367-293981-0003.flac", + "question": "When is 'think' first mentioned in the clip?", + "choice_a": "[0.36, 0.55]", + "choice_b": "[6.83, 7.28]", + "choice_c": "[7.91, 8.16]", + "choice_d": "[4.33, 4.47]", + "answer_gt": "[7.91, 8.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 310 + }, + { + "path": "3752-4944-0030.flac", + "question": "What is the duration of 'he's' being mentioned in the conversation?", + "choice_a": "[0.95, 1.05]", + "choice_b": "[0.44, 0.63]", + "choice_c": "[1.35, 1.54]", + "choice_d": "[1.83, 2.24]", + "answer_gt": "[1.35, 1.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 311 + }, + { + "path": "1993-147149-0018.flac", + "question": "Locate the segment where 'thrilled' appears most frequently.", + "choice_a": "[6.43, 6.54]", + "choice_b": "[8.09, 8.45]", + "choice_c": "[6.54, 7.09]", + "choice_d": "[7.09, 7.59]", + "answer_gt": "[8.09, 8.45]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 312 + }, + { + "path": "700-122867-0024.flac", + "question": "Listen for the occurrence of 'dismal' in the recording.", + "choice_a": "[1.59, 1.75]", + "choice_b": "[0.56, 0.97]", + "choice_c": "[1.75, 1.87]", + "choice_d": "[1.87, 1.94]", + "answer_gt": "[0.56, 0.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 313 + }, + { + "path": "8463-294825-0009.flac", + "question": "Discover when 'yet' is introduced in the conversation.", + "choice_a": "[0.22, 10.5]", + "choice_b": "[6.31, 6.40]", + "choice_c": "[6.05, 17.0]", + "choice_d": "[5.79, 5.97]", + "answer_gt": "[5.79, 5.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 314 + }, + { + "path": "2506-11278-0004.flac", + "question": "Listen to this audio clip; identify the time segment when 'and' is spoken.", + "choice_a": "[2.70, 3.48]", + "choice_b": "[4.51, 4.71]", + "choice_c": "[0.65, 1.54]", + "choice_d": "[2.17, 2.65]", + "answer_gt": "[4.51, 4.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 315 + }, + { + "path": "7641-96684-0008.flac", + "question": "What is the timestamp for the initial occurrence of 'he'?", + "choice_a": "[1.74, 1.93]", + "choice_b": "[2.42, 2.84]", + "choice_c": "[0.22, 0.35]", + "choice_d": "[1.06, 1.35]", + "answer_gt": "[0.22, 0.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 316 + }, + { + "path": "3660-6517-0015.flac", + "question": "Recognize the time when 'company's' is most clearly articulated.", + "choice_a": "[1.23, 1.26]", + "choice_b": "[1.68, 1.73]", + "choice_c": "[2.27, 2.64]", + "choice_d": "[1.59, 1.68]", + "answer_gt": "[2.27, 2.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 317 + }, + { + "path": "908-157963-0021.flac", + "question": "Note the time when 'or' is first mentioned in the audio.", + "choice_a": "[6.80, 6.96]", + "choice_b": "[1.70, 1.76]", + "choice_c": "[6.80, 6.96]", + "choice_d": "[4.61, 5.22]", + "answer_gt": "[1.70, 1.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 318 + }, + { + "path": "4153-186222-0021.flac", + "question": "Determine the point in the audio when 'to' is discussed.", + "choice_a": "[1.13, 1.20]", + "choice_b": "[0.89, 1.13]", + "choice_c": "[0.89, 1.13]", + "choice_d": "[0.89, 1.13]", + "answer_gt": "[1.13, 1.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 319 + }, + { + "path": "7697-105815-0018.flac", + "question": "At what point in the audio is 'what' clearly audible?", + "choice_a": "[0.95, 1.12]", + "choice_b": "[1.92, 2.10]", + "choice_c": "[0.31, 0.67]", + "choice_d": "[3.18, 3.32]", + "answer_gt": "[3.18, 3.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 320 + }, + { + "path": "5764-299665-0023.flac", + "question": "Determine the point in the audio when 'people' is repeated.", + "choice_a": "[2.37, 2.55]", + "choice_b": "[2.58, 2.64]", + "choice_c": "[0.55, 1.05]", + "choice_d": "[1.25, 1.78]", + "answer_gt": "[1.25, 1.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 321 + }, + { + "path": "2803-154328-0010.flac", + "question": "When does the speaker mention 'my'?", + "choice_a": "[2.97, 3.73]", + "choice_b": "[1.78, 2.26]", + "choice_c": "[2.78, 2.97]", + "choice_d": "[4.50, 4.88]", + "answer_gt": "[2.78, 2.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 322 + }, + { + "path": "3170-137482-0031.flac", + "question": "Highlight the time when 'unwilling' is mentioned.", + "choice_a": "[0.39, 0.80]", + "choice_b": "[1.94, 12.1]", + "choice_c": "[1.24, 1.40]", + "choice_d": "[0.80, 1.01]", + "answer_gt": "[0.39, 0.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 323 + }, + { + "path": "3915-98647-0031.flac", + "question": "Locate the part of the audio where 'the' is expressed.", + "choice_a": "[3.33, 3.49]", + "choice_b": "[0.45, 0.68]", + "choice_c": "[2.32, 2.77]", + "choice_d": "[0.95, 1.04]", + "answer_gt": "[0.95, 1.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 324 + }, + { + "path": "5895-34622-0021.flac", + "question": "Determine the precise moment 'what' in the audio's pace.", + "choice_a": "[0.62, 0.92]", + "choice_b": "[0.51, 0.62]", + "choice_c": "[0.32, 0.51]", + "choice_d": "[1.70, 1.87]", + "answer_gt": "[1.70, 1.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 325 + }, + { + "path": "8188-269290-0004.flac", + "question": "Highlight when 'nothing' is spoken.", + "choice_a": "[6.06, 6.28]", + "choice_b": "[0.56, 0.71]", + "choice_c": "[3.67, 3.94]", + "choice_d": "[3.23, 3.29]", + "answer_gt": "[6.06, 6.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 326 + }, + { + "path": "533-1066-0016.flac", + "question": "Recognize the time when 'know' is most clearly articulated.", + "choice_a": "[0.74, 0.98]", + "choice_b": "[0.98, 1.17]", + "choice_c": "[0.28, 0.42]", + "choice_d": "[0.28, 0.42]", + "answer_gt": "[0.74, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 327 + }, + { + "path": "2506-11278-0029.flac", + "question": "Identify the moment 'young' is mentioned in the audio.", + "choice_a": "[1.00, 1.52]", + "choice_b": "[1.52, 1.75]", + "choice_c": "[1.75, 1.95]", + "choice_d": "[0.56, 1.00]", + "answer_gt": "[1.75, 1.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 328 + }, + { + "path": "260-123440-0015.flac", + "question": "Determine the time when 'much' happens within the clip.", + "choice_a": "[4.45, 4.75]", + "choice_b": "[1.84, 2.32]", + "choice_c": "[4.81, 5.12]", + "choice_d": "[2.35, 2.62]", + "answer_gt": "[1.84, 2.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 329 + }, + { + "path": "700-122867-0025.flac", + "question": "Identify the portion of the clip where 'because' is most prominent.", + "choice_a": "[3.10, 3.35]", + "choice_b": "[3.83, 4.23]", + "choice_c": "[4.23, 4.69]", + "choice_d": "[2.82, 3.03]", + "answer_gt": "[3.10, 3.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 330 + }, + { + "path": "5142-36377-0003.flac", + "question": "When does 'and' occur in the audio?", + "choice_a": "[0.89, 1.07]", + "choice_b": "[1.26, 1.31]", + "choice_c": "[2.72, 2.90]", + "choice_d": "[2.16, 2.26]", + "answer_gt": "[2.16, 2.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 331 + }, + { + "path": "3536-23268-0000.flac", + "question": "Find the moment 'sir' is clearly audible in the audio.", + "choice_a": "[2.79, 2.90]", + "choice_b": "[3.07, 13.7]", + "choice_c": "[4.85, 15.2]", + "choice_d": "[0.46, 0.75]", + "answer_gt": "[0.46, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 332 + }, + { + "path": "3081-166546-0031.flac", + "question": "When is 'chair' first heard in the recording?", + "choice_a": "[1.34, 1.57]", + "choice_b": "[0.56, 0.87]", + "choice_c": "[8.93, 9.31]", + "choice_d": "[8.88, 8.93]", + "answer_gt": "[8.93, 9.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 333 + }, + { + "path": "1701-141759-0033.flac", + "question": "Note the time interval that includes the mention of 'always'.", + "choice_a": "[0.45, 0.76]", + "choice_b": "[1.74, 1.93]", + "choice_c": "[0.29, 0.39]", + "choice_d": "[0.76, 1.24]", + "answer_gt": "[0.45, 0.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 334 + }, + { + "path": "1580-141084-0013.flac", + "question": "Find the duration when 'the' occurs consecutively in the audio.", + "choice_a": "[0.88, 1.02]", + "choice_b": "[0.44, 0.57]", + "choice_c": "[0.57, 0.65]", + "choice_d": "[0.25, 0.44]", + "answer_gt": "[0.57, 0.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 335 + }, + { + "path": "7127-75946-0015.flac", + "question": "Determine the time when 'accompanied' happens within the clip.", + "choice_a": "[5.65, 6.18]", + "choice_b": "[4.47, 5.04]", + "choice_c": "[1.30, 1.46]", + "choice_d": "[3.69, 4.44]", + "answer_gt": "[4.47, 5.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 336 + }, + { + "path": "2961-961-0016.flac", + "question": "Listen for when 'them' is emphasized by the narrator.", + "choice_a": "[3.74, 3.86]", + "choice_b": "[2.23, 2.55]", + "choice_c": "[4.33, 4.45]", + "choice_d": "[1.93, 2.08]", + "answer_gt": "[1.93, 2.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 337 + }, + { + "path": "3752-4944-0047.flac", + "question": "Listen for the precise interval when 'will' is dominant.", + "choice_a": "[6.19, 6.34]", + "choice_b": "[0.16, 0.47]", + "choice_c": "[5.78, 5.90]", + "choice_d": "[2.95, 3.31]", + "answer_gt": "[5.78, 5.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 338 + }, + { + "path": "6938-70848-0020.flac", + "question": "When is 'will' first mentioned in the clip?", + "choice_a": "[2.59, 3.04]", + "choice_b": "[1.62, 2.02]", + "choice_c": "[1.47, 1.62]", + "choice_d": "[1.47, 1.62]", + "answer_gt": "[1.47, 1.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 339 + }, + { + "path": "7902-96594-0030.flac", + "question": "Determine the point in the audio when 'say' is discussed.", + "choice_a": "[5.24, 5.49]", + "choice_b": "[1.85, 1.98]", + "choice_c": "[0.98, 1.18]", + "choice_d": "[2.38, 2.55]", + "answer_gt": "[2.38, 2.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 340 + }, + { + "path": "1585-131718-0016.flac", + "question": "Identify the portion of the clip where 'addition' is most prominent.", + "choice_a": "[5.82, 6.45]", + "choice_b": "[5.11, 5.58]", + "choice_c": "[3.49, 3.57]", + "choice_d": "[1.13, 1.29]", + "answer_gt": "[5.11, 5.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 341 + }, + { + "path": "8297-275154-0023.flac", + "question": "Listen for when 'hear' is emphasized by the narrator.", + "choice_a": "[0.65, 0.83]", + "choice_b": "[0.83, 1.00]", + "choice_c": "[1.27, 1.49]", + "choice_d": "[1.00, 1.19]", + "answer_gt": "[0.83, 1.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 342 + }, + { + "path": "6432-63722-0036.flac", + "question": "Locate the segment where 'colonel' appears most frequently.", + "choice_a": "[2.64, 2.80]", + "choice_b": "[1.44, 1.50]", + "choice_c": "[3.91, 4.07]", + "choice_d": "[1.50, 1.92]", + "answer_gt": "[1.50, 1.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 343 + }, + { + "path": "6930-75918-0011.flac", + "question": "At what interval does 'said' appear in the audio?", + "choice_a": "[0.23, 0.37]", + "choice_b": "[2.18, 2.44]", + "choice_c": "[1.30, 1.47]", + "choice_d": "[0.48, 1.16]", + "answer_gt": "[2.18, 2.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 344 + }, + { + "path": "533-131556-0006.flac", + "question": "When is 'the' first mentioned in the clip?", + "choice_a": "[7.04, 7.48]", + "choice_b": "[3.52, 3.60]", + "choice_c": "[9.83, 9.91]", + "choice_d": "[4.16, 4.35]", + "answer_gt": "[9.83, 9.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 345 + }, + { + "path": "8131-117016-0042.flac", + "question": "Determine the precise moment 'finally' in the audio's pace.", + "choice_a": "[0.60, 0.67]", + "choice_b": "[1.58, 2.14]", + "choice_c": "[0.67, 1.12]", + "choice_d": "[0.47, 0.60]", + "answer_gt": "[1.58, 2.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 346 + }, + { + "path": "251-118436-0001.flac", + "question": "Listen for the occurrence of 'it' in the recording.", + "choice_a": "[0.37, 0.49]", + "choice_b": "[1.51, 1.82]", + "choice_c": "[1.88, 2.44]", + "choice_d": "[1.19, 1.31]", + "answer_gt": "[1.19, 1.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 347 + }, + { + "path": "1919-142785-0028.flac", + "question": "When does the speaker mention 'before'?", + "choice_a": "[1.30, 1.37]", + "choice_b": "[2.59, 2.79]", + "choice_c": "[3.77, 3.92]", + "choice_d": "[0.80, 1.15]", + "answer_gt": "[0.80, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 348 + }, + { + "path": "6241-66616-0020.flac", + "question": "Recognize the segment where 'and' is spoken by the speaker.", + "choice_a": "[1.14, 1.28]", + "choice_b": "[1.28, 1.78]", + "choice_c": "[1.95, 2.11]", + "choice_d": "[1.14, 1.28]", + "answer_gt": "[1.14, 1.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 349 + }, + { + "path": "1089-134686-0003.flac", + "question": "Pinpoint the passage where 'your' is heard in the background.", + "choice_a": "[1.25, 1.44]", + "choice_b": "[0.41, 0.68]", + "choice_c": "[1.85, 1.98]", + "choice_d": "[0.68, 1.25]", + "answer_gt": "[1.85, 1.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 350 + }, + { + "path": "7975-280057-0017.flac", + "question": "Listen for the occurrence of 'to' in the recording.", + "choice_a": "[1.55, 2.04]", + "choice_b": "[1.46, 1.55]", + "choice_c": "[1.10, 1.27]", + "choice_d": "[1.27, 1.46]", + "answer_gt": "[1.46, 1.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 351 + }, + { + "path": "6319-275224-0018.flac", + "question": "Recognize the time when 'hurried' is mentioned by the speaker.", + "choice_a": "[4.37, 4.66]", + "choice_b": "[2.34, 2.61]", + "choice_c": "[3.45, 3.72]", + "choice_d": "[2.97, 3.16]", + "answer_gt": "[4.37, 4.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 352 + }, + { + "path": "5484-24317-0022.flac", + "question": "Note the time when 'said' is first mentioned in the audio.", + "choice_a": "[3.13, 3.40]", + "choice_b": "[0.55, 0.71]", + "choice_c": "[2.85, 3.13]", + "choice_d": "[3.94, 4.33]", + "answer_gt": "[2.85, 3.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 353 + }, + { + "path": "6432-63722-0017.flac", + "question": "Listen for when 'about' is emphasized by the narrator.", + "choice_a": "[1.65, 1.86]", + "choice_b": "[2.15, 2.49]", + "choice_c": "[1.86, 2.15]", + "choice_d": "[1.45, 1.52]", + "answer_gt": "[2.15, 2.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 354 + }, + { + "path": "3729-6852-0018.flac", + "question": "In the audio track, when is 'a' first noticed?", + "choice_a": "[1.51, 1.56]", + "choice_b": "[1.43, 1.51]", + "choice_c": "[1.03, 1.43]", + "choice_d": "[1.56, 1.99]", + "answer_gt": "[1.51, 1.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 355 + }, + { + "path": "121-127105-0012.flac", + "question": "Spot the exact moment 'i' appears in the audio.", + "choice_a": "[2.32, 2.57]", + "choice_b": "[1.53, 1.89]", + "choice_c": "[0.50, 0.84]", + "choice_d": "[2.23, 2.32]", + "answer_gt": "[2.23, 2.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 356 + }, + { + "path": "6128-63241-0002.flac", + "question": "Identify when 'be' is heard in the audio.", + "choice_a": "[0.84, 1.04]", + "choice_b": "[0.42, 0.84]", + "choice_c": "[4.12, 4.27]", + "choice_d": "[4.58, 4.68]", + "answer_gt": "[4.58, 4.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 357 + }, + { + "path": "1701-141760-0044.flac", + "question": "Locate the part of the audio where 'was' is expressed.", + "choice_a": "[6.52, 6.82]", + "choice_b": "[2.62, 2.86]", + "choice_c": "[3.96, 4.20]", + "choice_d": "[4.44, 4.64]", + "answer_gt": "[4.44, 4.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 358 + }, + { + "path": "6267-65525-0006.flac", + "question": "Listen to this audio clip: when does 'her' occur?", + "choice_a": "[2.91, 3.42]", + "choice_b": "[5.18, 5.30]", + "choice_c": "[1.59, 1.96]", + "choice_d": "[0.97, 1.59]", + "answer_gt": "[5.18, 5.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 359 + }, + { + "path": "2428-83705-0012.flac", + "question": "Detect the point at which 'a' is spoken.", + "choice_a": "[3.39, 3.99]", + "choice_b": "[4.45, 4.51]", + "choice_c": "[5.28, 5.35]", + "choice_d": "[5.70, 5.98]", + "answer_gt": "[4.45, 4.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 360 + }, + { + "path": "7902-96592-0003.flac", + "question": "Listen to this audio clip; identify the time segment when 'down' is spoken.", + "choice_a": "[1.65, 1.99]", + "choice_b": "[8.90, 9.11]", + "choice_c": "[7.29, 7.51]", + "choice_d": "[0.23, 0.55]", + "answer_gt": "[1.65, 1.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 361 + }, + { + "path": "6313-66129-0020.flac", + "question": "Identify the time segment when 'all' is mentioned in the audio.", + "choice_a": "[2.28, 2.53]", + "choice_b": "[0.33, 0.40]", + "choice_c": "[3.14, 3.43]", + "choice_d": "[5.08, 5.48]", + "answer_gt": "[2.28, 2.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 362 + }, + { + "path": "5639-40744-0026.flac", + "question": "Identify when 'by' is heard in the audio.", + "choice_a": "[6.01, 6.32]", + "choice_b": "[4.33, 4.46]", + "choice_c": "[3.87, 4.18]", + "choice_d": "[7.71, 7.91]", + "answer_gt": "[7.71, 7.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 363 + }, + { + "path": "5142-33396-0053.flac", + "question": "Discover when 'our' is introduced in the conversation.", + "choice_a": "[1.19, 1.62]", + "choice_b": "[1.69, 1.96]", + "choice_c": "[2.11, 2.18]", + "choice_d": "[3.27, 3.39]", + "answer_gt": "[2.11, 2.18]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 364 + }, + { + "path": "2033-164914-0012.flac", + "question": "Ascertain the point at which 'linger' enters the discussion.", + "choice_a": "[1.69, 2.02]", + "choice_b": "[0.40, 0.78]", + "choice_c": "[1.52, 1.69]", + "choice_d": "[1.52, 1.69]", + "answer_gt": "[1.69, 2.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 365 + }, + { + "path": "7697-105815-0040.flac", + "question": "At what point in the audio is 'sir' clearly audible?", + "choice_a": "[1.31, 1.74]", + "choice_b": "[3.10, 3.21]", + "choice_c": "[0.42, 0.88]", + "choice_d": "[3.21, 3.27]", + "answer_gt": "[1.31, 1.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 366 + }, + { + "path": "3729-6852-0032.flac", + "question": "Highlight when 'have' is spoken.", + "choice_a": "[4.87, 5.19]", + "choice_b": "[1.28, 11.5]", + "choice_c": "[5.40, 5.74]", + "choice_d": "[3.97, 4.06]", + "answer_gt": "[4.87, 5.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 367 + }, + { + "path": "3575-170457-0044.flac", + "question": "Highlight when 'enjoyment' is spoken.", + "choice_a": "[2.52, 2.90]", + "choice_b": "[8.80, 8.92]", + "choice_c": "[7.29, 7.58]", + "choice_d": "[4.45, 4.94]", + "answer_gt": "[4.45, 4.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 368 + }, + { + "path": "174-50561-0010.flac", + "question": "Determine the point in the audio when 'leave' is discussed.", + "choice_a": "[1.90, 2.01]", + "choice_b": "[2.51, 2.82]", + "choice_c": "[6.45, 6.68]", + "choice_d": "[1.54, 1.60]", + "answer_gt": "[6.45, 6.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 369 + }, + { + "path": "5764-299665-0070.flac", + "question": "Find the duration when 'a' occurs consecutively in the audio.", + "choice_a": "[1.03, 1.09]", + "choice_b": "[0.78, 1.03]", + "choice_c": "[0.51, 0.78]", + "choice_d": "[1.09, 1.75]", + "answer_gt": "[1.03, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 370 + }, + { + "path": "8131-117017-0010.flac", + "question": "When in the audio does 'the' coincide with a specific event?", + "choice_a": "[1.73, 1.80]", + "choice_b": "[1.80, 2.04]", + "choice_c": "[2.69, 3.14]", + "choice_d": "[0.58, 0.65]", + "answer_gt": "[1.73, 1.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 371 + }, + { + "path": "1988-24833-0010.flac", + "question": "Locate the segment where 'he' appears most frequently.", + "choice_a": "[1.20, 1.76]", + "choice_b": "[1.14, 1.20]", + "choice_c": "[0.52, 0.98]", + "choice_d": "[0.52, 0.98]", + "answer_gt": "[1.14, 1.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 372 + }, + { + "path": "7176-92135-0045.flac", + "question": "When in the audio does 'left' coincide with a specific event?", + "choice_a": "[2.98, 3.12]", + "choice_b": "[0.45, 0.70]", + "choice_c": "[2.02, 2.48]", + "choice_d": "[3.12, 3.50]", + "answer_gt": "[3.12, 3.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 373 + }, + { + "path": "1221-135767-0004.flac", + "question": "When does 'in' occur in the audio?", + "choice_a": "[0.70, 1.15]", + "choice_b": "[5.90, 16.0]", + "choice_c": "[1.32, 1.75]", + "choice_d": "[7.04, 17.1]", + "answer_gt": "[5.90, 16.0]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 374 + }, + { + "path": "777-126732-0056.flac", + "question": "What is the timestamp for the initial occurrence of 'several'?", + "choice_a": "[5.65, 5.81]", + "choice_b": "[1.40, 1.51]", + "choice_c": "[7.76, 8.11]", + "choice_d": "[6.79, 7.11]", + "answer_gt": "[7.76, 8.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 375 + }, + { + "path": "3915-98647-0017.flac", + "question": "Recognize the time when 'rites' is most clearly articulated.", + "choice_a": "[2.36, 2.43]", + "choice_b": "[0.51, 0.71]", + "choice_c": "[4.38, 4.79]", + "choice_d": "[7.06, 7.40]", + "answer_gt": "[7.06, 7.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 376 + }, + { + "path": "8461-258277-0010.flac", + "question": "At what interval does 'is' appear in the audio?", + "choice_a": "[4.86, 5.00]", + "choice_b": "[5.73, 5.88]", + "choice_c": "[0.76, 0.91]", + "choice_d": "[1.65, 2.13]", + "answer_gt": "[4.86, 5.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 377 + }, + { + "path": "2414-159411-0020.flac", + "question": "Listen to this audio clip: when does 'patience' occur?", + "choice_a": "[8.05, 8.26]", + "choice_b": "[4.15, 4.58]", + "choice_c": "[5.12, 5.27]", + "choice_d": "[6.73, 7.36]", + "answer_gt": "[6.73, 7.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 378 + }, + { + "path": "777-126732-0060.flac", + "question": "When is 'time' first mentioned in the clip?", + "choice_a": "[0.58, 1.11]", + "choice_b": "[2.93, 3.35]", + "choice_c": "[1.79, 2.12]", + "choice_d": "[4.60, 5.02]", + "answer_gt": "[4.60, 5.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 379 + }, + { + "path": "6841-88294-0048.flac", + "question": "Pinpoint the exact moment 'kind' is repeated in the audio.", + "choice_a": "[1.41, 1.57]", + "choice_b": "[0.88, 0.94]", + "choice_c": "[0.97, 1.15]", + "choice_d": "[0.67, 0.88]", + "answer_gt": "[0.67, 0.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 380 + }, + { + "path": "2803-161169-0017.flac", + "question": "Identify the time segment when 'does' is mentioned in the audio.", + "choice_a": "[7.03, 7.36]", + "choice_b": "[4.69, 5.22]", + "choice_c": "[0.54, 1.12]", + "choice_d": "[4.38, 4.58]", + "answer_gt": "[4.38, 4.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 381 + }, + { + "path": "8254-84205-0010.flac", + "question": "When in the audio does 'stiff' coincide with a specific event?", + "choice_a": "[0.86, 1.06]", + "choice_b": "[0.49, 0.75]", + "choice_c": "[1.11, 1.31]", + "choice_d": "[1.31, 1.69]", + "answer_gt": "[1.31, 1.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 382 + }, + { + "path": "4992-23283-0000.flac", + "question": "Identify the time segment when 'remembrance' is mentioned in the audio.", + "choice_a": "[1.98, 2.17]", + "choice_b": "[1.22, 1.98]", + "choice_c": "[4.74, 5.39]", + "choice_d": "[5.39, 5.51]", + "answer_gt": "[4.74, 5.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 383 + }, + { + "path": "2277-149896-0026.flac", + "question": "Pinpoint the exact moment 'up' is repeated in the audio.", + "choice_a": "[2.70, 2.87]", + "choice_b": "[0.97, 1.11]", + "choice_c": "[2.44, 2.70]", + "choice_d": "[0.58, 0.97]", + "answer_gt": "[2.70, 2.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 384 + }, + { + "path": "3081-166546-0027.flac", + "question": "Determine the point in the audio when 'a' is repeated.", + "choice_a": "[1.24, 1.62]", + "choice_b": "[1.16, 1.24]", + "choice_c": "[1.16, 1.24]", + "choice_d": "[1.62, 1.99]", + "answer_gt": "[1.16, 1.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 385 + }, + { + "path": "7902-96591-0016.flac", + "question": "Listen for when 'the' is emphasized by the narrator.", + "choice_a": "[4.65, 4.80]", + "choice_b": "[0.30, 0.36]", + "choice_c": "[1.89, 2.02]", + "choice_d": "[0.87, 1.33]", + "answer_gt": "[0.30, 0.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 386 + }, + { + "path": "4572-112383-0002.flac", + "question": "Pinpoint the exact moment 'the' is repeated in the audio.", + "choice_a": "[2.22, 2.37]", + "choice_b": "[3.65, 3.88]", + "choice_c": "[1.48, 1.93]", + "choice_d": "[1.37, 1.48]", + "answer_gt": "[1.37, 1.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 387 + }, + { + "path": "3538-163624-0025.flac", + "question": "Find out when the speaker refers to 'ridden' during the audio.", + "choice_a": "[1.23, 1.36]", + "choice_b": "[1.76, 2.06]", + "choice_c": "[3.24, 3.44]", + "choice_d": "[2.06, 2.29]", + "answer_gt": "[1.76, 2.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 388 + }, + { + "path": "4852-28312-0030.flac", + "question": "Listen to this audio clip; identify the time segment when 'inn' is spoken.", + "choice_a": "[0.80, 1.03]", + "choice_b": "[1.86, 2.22]", + "choice_c": "[0.32, 0.60]", + "choice_d": "[1.03, 1.38]", + "answer_gt": "[1.03, 1.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 389 + }, + { + "path": "8173-294714-0015.flac", + "question": "When in the audio does 'such' coincide with a specific event?", + "choice_a": "[1.29, 1.46]", + "choice_b": "[0.40, 0.70]", + "choice_c": "[0.70, 0.75]", + "choice_d": "[0.27, 0.40]", + "answer_gt": "[0.40, 0.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 390 + }, + { + "path": "1630-141772-0005.flac", + "question": "Note the time when 'lifted' is first mentioned in the audio.", + "choice_a": "[6.17, 6.55]", + "choice_b": "[9.65, 9.84]", + "choice_c": "[2.31, 2.62]", + "choice_d": "[2.88, 2.99]", + "answer_gt": "[6.17, 6.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 391 + }, + { + "path": "1580-141083-0046.flac", + "question": "Locate the segment where 'done' appears most frequently.", + "choice_a": "[1.34, 1.52]", + "choice_b": "[1.58, 1.91]", + "choice_c": "[1.52, 1.58]", + "choice_d": "[2.14, 2.21]", + "answer_gt": "[1.34, 1.52]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 392 + }, + { + "path": "6432-63723-0047.flac", + "question": "Listen to this audio clip; identify the time segment when 'be' is spoken.", + "choice_a": "[3.00, 3.06]", + "choice_b": "[3.06, 3.19]", + "choice_c": "[0.81, 0.98]", + "choice_d": "[3.19, 3.45]", + "answer_gt": "[3.06, 3.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 393 + }, + { + "path": "3915-57461-0023.flac", + "question": "Locate the segment where 'therefore' appears most frequently.", + "choice_a": "[2.23, 2.38]", + "choice_b": "[1.10, 1.60]", + "choice_c": "[1.60, 1.72]", + "choice_d": "[1.72, 2.23]", + "answer_gt": "[1.10, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 394 + }, + { + "path": "7697-245712-0004.flac", + "question": "Listen to this audio clip; identify the time segment when 'proved' is spoken.", + "choice_a": "[2.09, 2.49]", + "choice_b": "[2.62, 2.94]", + "choice_c": "[1.35, 1.42]", + "choice_d": "[1.42, 1.79]", + "answer_gt": "[1.42, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 395 + }, + { + "path": "1998-29454-0014.flac", + "question": "When in the audio does 'fair' coincide with a specific event?", + "choice_a": "[0.87, 1.06]", + "choice_b": "[0.23, 0.29]", + "choice_c": "[1.06, 1.39]", + "choice_d": "[0.42, 0.73]", + "answer_gt": "[0.87, 1.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 396 + }, + { + "path": "3660-172182-0009.flac", + "question": "At what point in the audio is 'then' clearly audible?", + "choice_a": "[2.02, 2.24]", + "choice_b": "[0.56, 0.69]", + "choice_c": "[1.87, 2.02]", + "choice_d": "[0.83, 1.38]", + "answer_gt": "[0.56, 0.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 397 + }, + { + "path": "1919-142785-0035.flac", + "question": "Pinpoint the exact moment 'sufficient' appears.", + "choice_a": "[1.02, 1.14]", + "choice_b": "[2.29, 2.71]", + "choice_c": "[0.46, 1.02]", + "choice_d": "[1.14, 1.55]", + "answer_gt": "[0.46, 1.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 398 + }, + { + "path": "2414-159411-0033.flac", + "question": "Note the time interval that includes the mention of 'you'.", + "choice_a": "[3.73, 3.90]", + "choice_b": "[4.77, 5.36]", + "choice_c": "[2.18, 2.49]", + "choice_d": "[5.48, 5.57]", + "answer_gt": "[5.48, 5.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 399 + }, + { + "path": "3536-23268-0016.flac", + "question": "In the audio track, when is 'thoughts' first noticed?", + "choice_a": "[3.12, 3.46]", + "choice_b": "[0.63, 0.98]", + "choice_c": "[6.54, 6.85]", + "choice_d": "[8.65, 8.88]", + "answer_gt": "[3.12, 3.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 400 + }, + { + "path": "1630-141772-0001.flac", + "question": "Ascertain the exact time 'lacerating' is mentioned in the audio.", + "choice_a": "[5.01, 5.78]", + "choice_b": "[1.08, 1.46]", + "choice_c": "[3.41, 3.87]", + "choice_d": "[0.86, 1.03]", + "answer_gt": "[5.01, 5.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 401 + }, + { + "path": "3528-168669-0098.flac", + "question": "Recognize the time when 'to' is mentioned by the speaker.", + "choice_a": "[2.25, 2.32]", + "choice_b": "[1.80, 1.94]", + "choice_c": "[1.94, 2.25]", + "choice_d": "[0.54, 0.60]", + "answer_gt": "[2.25, 2.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 402 + }, + { + "path": "7176-92135-0033.flac", + "question": "At what interval does 'the' appear in the audio?", + "choice_a": "[1.37, 1.42]", + "choice_b": "[1.37, 1.42]", + "choice_c": "[0.88, 1.26]", + "choice_d": "[0.71, 0.88]", + "answer_gt": "[1.37, 1.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 403 + }, + { + "path": "1272-135031-0020.flac", + "question": "What is the timestamp for the initial occurrence of 'don't'?", + "choice_a": "[0.56, 0.77]", + "choice_b": "[2.94, 3.17]", + "choice_c": "[1.55, 1.84]", + "choice_d": "[3.54, 3.63]", + "answer_gt": "[0.56, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 404 + }, + { + "path": "251-137823-0002.flac", + "question": "Identify the time segment when 'your' is mentioned in the audio.", + "choice_a": "[1.04, 1.15]", + "choice_b": "[0.24, 0.45]", + "choice_c": "[1.66, 2.15]", + "choice_d": "[2.15, 2.33]", + "answer_gt": "[1.04, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 405 + }, + { + "path": "3997-180294-0011.flac", + "question": "Detect the point at which 'they' is spoken.", + "choice_a": "[0.90, 0.98]", + "choice_b": "[0.23, 0.39]", + "choice_c": "[0.98, 1.63]", + "choice_d": "[1.63, 1.86]", + "answer_gt": "[0.23, 0.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 406 + }, + { + "path": "7127-75947-0014.flac", + "question": "Note the time interval that includes the mention of 'me'.", + "choice_a": "[6.03, 6.35]", + "choice_b": "[5.61, 6.03]", + "choice_c": "[0.69, 0.92]", + "choice_d": "[5.08, 5.44]", + "answer_gt": "[0.69, 0.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 407 + }, + { + "path": "3080-5040-0006.flac", + "question": "Spot the exact moment 'for' appears in the audio.", + "choice_a": "[1.84, 1.99]", + "choice_b": "[4.12, 4.27]", + "choice_c": "[2.79, 2.91]", + "choice_d": "[6.18, 6.85]", + "answer_gt": "[1.84, 1.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 408 + }, + { + "path": "700-122867-0039.flac", + "question": "Locate the segment where 'call' is faintly heard.", + "choice_a": "[0.41, 0.49]", + "choice_b": "[0.60, 0.80]", + "choice_c": "[2.48, 2.60]", + "choice_d": "[0.49, 0.60]", + "answer_gt": "[0.60, 0.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 409 + }, + { + "path": "4515-11057-0059.flac", + "question": "Highlight the time when 'occurred' is mentioned.", + "choice_a": "[1.10, 1.34]", + "choice_b": "[1.76, 1.88]", + "choice_c": "[1.34, 1.76]", + "choice_d": "[0.61, 0.79]", + "answer_gt": "[1.34, 1.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 410 + }, + { + "path": "6128-63240-0013.flac", + "question": "When in the audio does 'attractive' coincide with a specific event?", + "choice_a": "[0.90, 1.51]", + "choice_b": "[0.53, 0.71]", + "choice_c": "[1.67, 2.57]", + "choice_d": "[1.51, 1.67]", + "answer_gt": "[0.90, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 411 + }, + { + "path": "3853-163249-0005.flac", + "question": "Listen to this audio clip; identify the time segment when 'mind' is spoken.", + "choice_a": "[2.32, 2.39]", + "choice_b": "[2.11, 2.32]", + "choice_c": "[8.64, 8.96]", + "choice_d": "[2.81, 3.51]", + "answer_gt": "[8.64, 8.96]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 412 + }, + { + "path": "2035-147960-0012.flac", + "question": "Recognize the time when 'my' is most clearly articulated.", + "choice_a": "[2.65, 3.11]", + "choice_b": "[1.76, 1.90]", + "choice_c": "[1.58, 1.76]", + "choice_d": "[1.08, 1.16]", + "answer_gt": "[1.76, 1.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 413 + }, + { + "path": "1988-24833-0028.flac", + "question": "Highlight when 'you' is spoken.", + "choice_a": "[1.41, 1.84]", + "choice_b": "[0.94, 1.24]", + "choice_c": "[1.84, 2.12]", + "choice_d": "[0.54, 0.83]", + "answer_gt": "[0.94, 1.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 414 + }, + { + "path": "672-122797-0069.flac", + "question": "Discover when 'in' is introduced in the conversation.", + "choice_a": "[0.53, 0.64]", + "choice_b": "[1.68, 1.79]", + "choice_c": "[0.80, 0.90]", + "choice_d": "[3.73, 3.94]", + "answer_gt": "[0.80, 0.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 415 + }, + { + "path": "4515-11057-0094.flac", + "question": "What is the timestamp for when 'what' occurs?", + "choice_a": "[1.51, 11.6]", + "choice_b": "[0.62, 0.82]", + "choice_c": "[0.22, 0.62]", + "choice_d": "[4.76, 4.86]", + "answer_gt": "[1.51, 11.6]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 416 + }, + { + "path": "5536-43363-0015.flac", + "question": "Identify the moment 'repeated' is mentioned in the audio.", + "choice_a": "[0.75, 1.10]", + "choice_b": "[1.20, 1.72]", + "choice_c": "[3.06, 3.41]", + "choice_d": "[0.27, 0.54]", + "answer_gt": "[1.20, 1.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 417 + }, + { + "path": "6319-275224-0006.flac", + "question": "Establish when 'stupid' is first mentioned in the clip.", + "choice_a": "[5.51, 6.00]", + "choice_b": "[0.95, 1.02]", + "choice_c": "[5.12, 15.7]", + "choice_d": "[2.81, 3.32]", + "answer_gt": "[5.12, 15.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 418 + }, + { + "path": "6841-88291-0018.flac", + "question": "When does the audio feature the sound of 'thus'?", + "choice_a": "[0.48, 0.59]", + "choice_b": "[0.59, 0.89]", + "choice_c": "[1.05, 1.44]", + "choice_d": "[0.31, 0.48]", + "answer_gt": "[0.31, 0.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 419 + }, + { + "path": "5849-50873-0008.flac", + "question": "Identify the moment 'the' is mentioned in the audio.", + "choice_a": "[2.65, 2.91]", + "choice_b": "[0.95, 1.23]", + "choice_c": "[3.75, 4.23]", + "choice_d": "[1.65, 2.09]", + "answer_gt": "[0.95, 1.23]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 420 + }, + { + "path": "2428-83699-0026.flac", + "question": "Highlight when 'be' is spoken.", + "choice_a": "[0.58, 0.98]", + "choice_b": "[0.58, 0.98]", + "choice_c": "[1.89, 2.07]", + "choice_d": "[0.28, 0.43]", + "answer_gt": "[0.58, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 421 + }, + { + "path": "8254-84205-0004.flac", + "question": "Locate the segment where 'was' appears most frequently.", + "choice_a": "[1.97, 2.06]", + "choice_b": "[0.96, 1.13]", + "choice_c": "[0.56, 0.79]", + "choice_d": "[4.34, 4.45]", + "answer_gt": "[0.96, 1.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 422 + }, + { + "path": "5484-24318-0035.flac", + "question": "Listen to this audio clip: when does 'walk' occur?", + "choice_a": "[2.66, 2.99]", + "choice_b": "[1.53, 1.62]", + "choice_c": "[3.66, 4.02]", + "choice_d": "[1.53, 1.62]", + "answer_gt": "[3.66, 4.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 423 + }, + { + "path": "6128-63241-0010.flac", + "question": "Ascertain the point at which 'that' enters the discussion.", + "choice_a": "[0.52, 0.70]", + "choice_b": "[1.72, 1.77]", + "choice_c": "[5.15, 5.35]", + "choice_d": "[3.46, 3.63]", + "answer_gt": "[3.46, 3.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 424 + }, + { + "path": "6345-93302-0020.flac", + "question": "Determine the point in the audio when 'can' is discussed.", + "choice_a": "[1.63, 1.82]", + "choice_b": "[1.82, 1.97]", + "choice_c": "[0.28, 0.66]", + "choice_d": "[1.97, 2.07]", + "answer_gt": "[1.82, 1.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 425 + }, + { + "path": "2078-142845-0036.flac", + "question": "Determine the point in the audio when 'twenty' is repeated.", + "choice_a": "[2.49, 2.87]", + "choice_b": "[1.67, 2.49]", + "choice_c": "[0.84, 1.52]", + "choice_d": "[0.84, 1.52]", + "answer_gt": "[2.49, 2.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 426 + }, + { + "path": "8131-117029-0016.flac", + "question": "Pinpoint the exact moment 'can't' appears.", + "choice_a": "[0.89, 1.19]", + "choice_b": "[0.59, 0.89]", + "choice_c": "[0.48, 0.59]", + "choice_d": "[0.48, 0.59]", + "answer_gt": "[0.59, 0.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 427 + }, + { + "path": "1919-142785-0052.flac", + "question": "When in the audio does 'and' coincide with a specific event?", + "choice_a": "[1.78, 2.18]", + "choice_b": "[0.54, 1.21]", + "choice_c": "[0.54, 1.21]", + "choice_d": "[2.18, 2.28]", + "answer_gt": "[2.18, 2.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 428 + }, + { + "path": "5543-27761-0009.flac", + "question": "Listen for the occurrence of 'came' in the recording.", + "choice_a": "[1.71, 2.00]", + "choice_b": "[2.36, 2.62]", + "choice_c": "[2.00, 2.23]", + "choice_d": "[1.41, 1.52]", + "answer_gt": "[2.00, 2.23]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 429 + }, + { + "path": "1585-157660-0006.flac", + "question": "Pinpoint the exact moment 'and' is repeated in the audio.", + "choice_a": "[1.30, 11.3]", + "choice_b": "[1.04, 1.52]", + "choice_c": "[0.93, 1.04]", + "choice_d": "[5.72, 5.98]", + "answer_gt": "[0.93, 1.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 430 + }, + { + "path": "8455-210777-0047.flac", + "question": "Highlight when 'you' is spoken.", + "choice_a": "[0.72, 1.16]", + "choice_b": "[1.16, 1.23]", + "choice_c": "[1.68, 1.84]", + "choice_d": "[0.58, 0.72]", + "answer_gt": "[0.58, 0.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 431 + }, + { + "path": "6123-59186-0026.flac", + "question": "Identify the moment 'each' is mentioned in the audio.", + "choice_a": "[6.08, 6.20]", + "choice_b": "[4.37, 4.68]", + "choice_c": "[2.76, 3.10]", + "choice_d": "[3.10, 3.18]", + "answer_gt": "[4.37, 4.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 432 + }, + { + "path": "700-122866-0033.flac", + "question": "When does 'girls' occur in the audio?", + "choice_a": "[0.43, 0.77]", + "choice_b": "[4.65, 4.74]", + "choice_c": "[4.38, 4.65]", + "choice_d": "[4.16, 4.29]", + "answer_gt": "[4.38, 4.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 433 + }, + { + "path": "5849-50873-0013.flac", + "question": "When does 'was' occur in the audio?", + "choice_a": "[0.37, 0.55]", + "choice_b": "[1.03, 1.16]", + "choice_c": "[1.16, 1.58]", + "choice_d": "[0.37, 0.55]", + "answer_gt": "[1.03, 1.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 434 + }, + { + "path": "237-126133-0001.flac", + "question": "Pinpoint the exact moment 'of' appears.", + "choice_a": "[5.21, 5.31]", + "choice_b": "[1.32, 1.51]", + "choice_c": "[2.57, 2.87]", + "choice_d": "[0.48, 0.77]", + "answer_gt": "[5.21, 5.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 435 + }, + { + "path": "3660-172183-0003.flac", + "question": "Find the moment 'confession' is clearly audible in the audio.", + "choice_a": "[3.74, 3.85]", + "choice_b": "[1.52, 2.12]", + "choice_c": "[0.54, 0.80]", + "choice_d": "[1.11, 1.33]", + "answer_gt": "[1.52, 2.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 436 + }, + { + "path": "6432-63723-0040.flac", + "question": "At what time interval does 'grew' occur in the audio clip?", + "choice_a": "[0.92, 1.40]", + "choice_b": "[0.71, 0.92]", + "choice_c": "[2.30, 2.65]", + "choice_d": "[3.64, 3.79]", + "answer_gt": "[3.64, 3.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 437 + }, + { + "path": "652-130726-0030.flac", + "question": "What is the timestamp for when 'and' occurs?", + "choice_a": "[0.71, 0.96]", + "choice_b": "[1.64, 1.83]", + "choice_c": "[0.96, 1.09]", + "choice_d": "[0.96, 1.09]", + "answer_gt": "[0.96, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 438 + }, + { + "path": "6070-63485-0015.flac", + "question": "Ascertain the exact time 'the' is mentioned in the audio.", + "choice_a": "[0.75, 1.07]", + "choice_b": "[1.14, 1.47]", + "choice_c": "[1.07, 1.14]", + "choice_d": "[0.70, 0.75]", + "answer_gt": "[0.70, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 439 + }, + { + "path": "2428-83699-0000.flac", + "question": "When in the audio does 'my' coincide with a specific event?", + "choice_a": "[7.10, 7.37]", + "choice_b": "[5.94, 6.10]", + "choice_c": "[7.83, 8.24]", + "choice_d": "[0.51, 0.60]", + "answer_gt": "[5.94, 6.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 440 + }, + { + "path": "6267-53049-0026.flac", + "question": "Find out when the speaker refers to 'her' during the audio.", + "choice_a": "[2.78, 3.22]", + "choice_b": "[9.49, 9.67]", + "choice_c": "[4.02, 4.25]", + "choice_d": "[0.89, 1.04]", + "answer_gt": "[4.02, 4.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 441 + }, + { + "path": "1688-142285-0051.flac", + "question": "Identify the time segment when 'sharpness' is mentioned in the audio.", + "choice_a": "[2.22, 2.79]", + "choice_b": "[0.52, 1.13]", + "choice_c": "[5.34, 5.70]", + "choice_d": "[1.28, 1.41]", + "answer_gt": "[0.52, 1.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 442 + }, + { + "path": "3170-137482-0017.flac", + "question": "At what interval does 'had' appear in the audio?", + "choice_a": "[2.21, 12.5]", + "choice_b": "[9.72, 10.3]", + "choice_c": "[2.05, 2.16]", + "choice_d": "[2.47, 2.66]", + "answer_gt": "[2.05, 2.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 443 + }, + { + "path": "6295-64301-0013.flac", + "question": "Recognize the time when 'and' is mentioned by the speaker.", + "choice_a": "[3.87, 4.24]", + "choice_b": "[2.74, 2.90]", + "choice_c": "[2.97, 3.73]", + "choice_d": "[4.24, 4.37]", + "answer_gt": "[2.74, 2.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 444 + }, + { + "path": "4077-13754-0014.flac", + "question": "At what time interval does 'the' occur in the audio clip?", + "choice_a": "[7.96, 8.12]", + "choice_b": "[8.20, 8.55]", + "choice_c": "[0.42, 0.49]", + "choice_d": "[6.29, 6.77]", + "answer_gt": "[0.42, 0.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 445 + }, + { + "path": "1701-141760-0047.flac", + "question": "Listen to this audio clip: when does 'our' occur?", + "choice_a": "[0.34, 0.54]", + "choice_b": "[2.85, 3.09]", + "choice_c": "[1.07, 1.14]", + "choice_d": "[1.01, 1.07]", + "answer_gt": "[0.34, 0.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 446 + }, + { + "path": "2094-142345-0059.flac", + "question": "Recognize the time when 'believing' is mentioned by the speaker.", + "choice_a": "[5.53, 6.01]", + "choice_b": "[6.75, 6.86]", + "choice_c": "[2.65, 3.11]", + "choice_d": "[0.78, 0.99]", + "answer_gt": "[2.65, 3.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 447 + }, + { + "path": "7601-175351-0004.flac", + "question": "Locate the segment where 'stove' is faintly heard.", + "choice_a": "[0.74, 1.25]", + "choice_b": "[1.38, 1.89]", + "choice_c": "[1.25, 1.38]", + "choice_d": "[3.18, 3.54]", + "answer_gt": "[1.38, 1.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 448 + }, + { + "path": "6467-94831-0009.flac", + "question": "Recognize the segment where 'come' is spoken by the speaker.", + "choice_a": "[1.09, 1.46]", + "choice_b": "[1.58, 1.75]", + "choice_c": "[0.87, 1.09]", + "choice_d": "[0.62, 0.87]", + "answer_gt": "[0.87, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 449 + }, + { + "path": "6841-88294-0021.flac", + "question": "When does the speaker mention 'so'?", + "choice_a": "[0.81, 0.87]", + "choice_b": "[1.15, 1.23]", + "choice_c": "[0.51, 0.81]", + "choice_d": "[0.87, 1.15]", + "answer_gt": "[0.87, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 450 + }, + { + "path": "3660-6517-0028.flac", + "question": "Detect the point at which 'of' is spoken.", + "choice_a": "[0.97, 1.25]", + "choice_b": "[1.25, 1.37]", + "choice_c": "[0.91, 0.97]", + "choice_d": "[1.43, 1.75]", + "answer_gt": "[1.25, 1.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 451 + }, + { + "path": "7021-85628-0004.flac", + "question": "At what point in the audio is 'why' clearly audible?", + "choice_a": "[1.44, 1.75]", + "choice_b": "[0.46, 1.00]", + "choice_c": "[1.27, 1.44]", + "choice_d": "[1.75, 1.98]", + "answer_gt": "[1.27, 1.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 452 + }, + { + "path": "4323-55228-0034.flac", + "question": "Determine the point in the audio when 'silence' is discussed.", + "choice_a": "[0.41, 0.51]", + "choice_b": "[0.51, 0.63]", + "choice_c": "[1.72, 1.87]", + "choice_d": "[0.89, 1.48]", + "answer_gt": "[0.89, 1.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 453 + }, + { + "path": "1998-29454-0041.flac", + "question": "Detect the point at which 'you're' is spoken.", + "choice_a": "[0.30, 0.50]", + "choice_b": "[0.50, 0.60]", + "choice_c": "[0.50, 0.60]", + "choice_d": "[1.00, 1.08]", + "answer_gt": "[0.50, 0.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 454 + }, + { + "path": "8461-258277-0008.flac", + "question": "Note the time interval that includes the mention of 'chest'.", + "choice_a": "[4.97, 5.08]", + "choice_b": "[4.45, 4.52]", + "choice_c": "[0.02, 10.5]", + "choice_d": "[0.71, 11.0]", + "answer_gt": "[0.71, 11.0]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 455 + }, + { + "path": "2277-149896-0017.flac", + "question": "Ascertain the point at which 'returned' enters the discussion.", + "choice_a": "[2.57, 2.70]", + "choice_b": "[1.15, 1.25]", + "choice_c": "[2.70, 2.83]", + "choice_d": "[2.89, 3.37]", + "answer_gt": "[2.89, 3.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 456 + }, + { + "path": "1580-141083-0052.flac", + "question": "Pinpoint the passage where 'oh' is heard in the background.", + "choice_a": "[1.20, 1.44]", + "choice_b": "[1.93, 2.06]", + "choice_c": "[2.06, 2.55]", + "choice_d": "[0.33, 0.84]", + "answer_gt": "[0.33, 0.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 457 + }, + { + "path": "5849-50873-0014.flac", + "question": "Recognize the segment where 'yes' is spoken by the speaker.", + "choice_a": "[0.47, 0.84]", + "choice_b": "[0.22, 0.47]", + "choice_c": "[0.84, 0.91]", + "choice_d": "[2.52, 2.59]", + "answer_gt": "[0.22, 0.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 458 + }, + { + "path": "4572-112375-0012.flac", + "question": "Find the duration when 'and' occurs consecutively in the audio.", + "choice_a": "[2.20, 2.37]", + "choice_b": "[2.37, 2.88]", + "choice_c": "[1.76, 2.14]", + "choice_d": "[4.71, 4.88]", + "answer_gt": "[2.20, 2.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 459 + }, + { + "path": "8297-275154-0011.flac", + "question": "At what time interval does 'down' occur in the audio clip?", + "choice_a": "[3.19, 3.31]", + "choice_b": "[3.31, 3.69]", + "choice_c": "[0.97, 1.29]", + "choice_d": "[0.71, 0.97]", + "answer_gt": "[0.97, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 460 + }, + { + "path": "5543-27761-0041.flac", + "question": "Determine the point in the audio when 'i'm' is repeated.", + "choice_a": "[2.57, 3.10]", + "choice_b": "[0.72, 1.10]", + "choice_c": "[1.13, 1.35]", + "choice_d": "[4.64, 5.21]", + "answer_gt": "[1.13, 1.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 461 + }, + { + "path": "8254-84205-0044.flac", + "question": "Determine the precise moment 'we're' in the audio's pace.", + "choice_a": "[1.06, 1.46]", + "choice_b": "[1.96, 2.02]", + "choice_c": "[2.76, 2.81]", + "choice_d": "[2.34, 2.44]", + "answer_gt": "[2.34, 2.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 462 + }, + { + "path": "3663-172005-0006.flac", + "question": "At what time interval does 'doe' occur in the audio clip?", + "choice_a": "[0.56, 0.91]", + "choice_b": "[0.48, 0.56]", + "choice_c": "[1.31, 1.73]", + "choice_d": "[0.91, 1.20]", + "answer_gt": "[0.56, 0.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 463 + }, + { + "path": "4852-28311-0019.flac", + "question": "When does 'his' occur in the audio?", + "choice_a": "[0.84, 0.89]", + "choice_b": "[1.24, 1.42]", + "choice_c": "[1.82, 2.14]", + "choice_d": "[1.42, 1.53]", + "answer_gt": "[1.42, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 464 + }, + { + "path": "5849-50873-0003.flac", + "question": "Highlight the time when 'refused' is mentioned.", + "choice_a": "[0.52, 0.76]", + "choice_b": "[2.00, 2.10]", + "choice_c": "[2.10, 2.46]", + "choice_d": "[1.47, 2.00]", + "answer_gt": "[1.47, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 465 + }, + { + "path": "5536-43359-0000.flac", + "question": "Identify the moment 'an' is mentioned in the audio.", + "choice_a": "[0.56, 0.68]", + "choice_b": "[1.98, 2.05]", + "choice_c": "[0.39, 0.56]", + "choice_d": "[0.68, 0.97]", + "answer_gt": "[1.98, 2.05]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 466 + }, + { + "path": "1686-142278-0009.flac", + "question": "Listen to this audio clip; identify the time segment when 'want' is spoken.", + "choice_a": "[0.64, 0.70]", + "choice_b": "[0.43, 0.64]", + "choice_c": "[0.39, 0.43]", + "choice_d": "[0.70, 0.99]", + "answer_gt": "[0.43, 0.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 467 + }, + { + "path": "1919-142785-0002.flac", + "question": "Listen to this audio clip; identify the time segment when 'valuable' is spoken.", + "choice_a": "[0.32, 0.93]", + "choice_b": "[4.84, 5.51]", + "choice_c": "[1.36, 1.83]", + "choice_d": "[1.03, 1.36]", + "answer_gt": "[1.36, 1.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 468 + }, + { + "path": "7975-280076-0019.flac", + "question": "What is the timestamp for when 'them' occurs?", + "choice_a": "[2.08, 2.31]", + "choice_b": "[1.81, 2.08]", + "choice_c": "[2.45, 2.85]", + "choice_d": "[0.98, 1.10]", + "answer_gt": "[2.08, 2.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 469 + }, + { + "path": "5639-40744-0018.flac", + "question": "Determine the point in the audio when 'doubt' is discussed.", + "choice_a": "[3.28, 13.5]", + "choice_b": "[9.48, 9.61]", + "choice_c": "[8.94, 9.34]", + "choice_d": "[7.38, 7.72]", + "answer_gt": "[7.38, 7.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 470 + }, + { + "path": "5639-40744-0038.flac", + "question": "Note the time when 'gladness' is first mentioned in the audio.", + "choice_a": "[5.48, 5.67]", + "choice_b": "[6.62, 6.77]", + "choice_c": "[2.11, 12.6]", + "choice_d": "[2.75, 2.91]", + "answer_gt": "[2.11, 12.6]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 471 + }, + { + "path": "1688-142285-0013.flac", + "question": "Detect the point at which 'missus' is spoken.", + "choice_a": "[5.35, 5.47]", + "choice_b": "[4.02, 4.20]", + "choice_c": "[0.64, 1.07]", + "choice_d": "[1.28, 1.44]", + "answer_gt": "[4.02, 4.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 472 + }, + { + "path": "6319-275224-0020.flac", + "question": "Listen for the precise interval when 'we' is dominant.", + "choice_a": "[1.89, 1.96]", + "choice_b": "[7.70, 7.85]", + "choice_c": "[5.53, 5.69]", + "choice_d": "[0.38, 10.5]", + "answer_gt": "[7.70, 7.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 473 + }, + { + "path": "5543-27761-0038.flac", + "question": "Pinpoint the exact moment 'why' appears.", + "choice_a": "[1.59, 1.75]", + "choice_b": "[0.52, 0.81]", + "choice_c": "[0.81, 1.10]", + "choice_d": "[0.81, 1.10]", + "answer_gt": "[0.52, 0.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 474 + }, + { + "path": "4153-186223-0010.flac", + "question": "In the audio track, when is 'boy' first noticed?", + "choice_a": "[0.93, 1.32]", + "choice_b": "[2.37, 2.53]", + "choice_c": "[2.59, 2.90]", + "choice_d": "[3.38, 3.75]", + "answer_gt": "[2.59, 2.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 475 + }, + { + "path": "2277-149874-0013.flac", + "question": "Recognize the segment where 'it' is spoken by the speaker.", + "choice_a": "[2.15, 2.37]", + "choice_b": "[0.57, 0.76]", + "choice_c": "[1.77, 1.90]", + "choice_d": "[1.90, 2.15]", + "answer_gt": "[1.77, 1.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 476 + }, + { + "path": "2609-156975-0011.flac", + "question": "Determine the point in the audio when 'he' is discussed.", + "choice_a": "[9.46, 9.59]", + "choice_b": "[8.05, 8.29]", + "choice_c": "[3.62, 3.67]", + "choice_d": "[0.44, 0.50]", + "answer_gt": "[3.62, 3.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 477 + }, + { + "path": "533-131556-0013.flac", + "question": "Recognize the time when 'into' is mentioned by the speaker.", + "choice_a": "[6.54, 6.64]", + "choice_b": "[1.11, 1.38]", + "choice_c": "[2.21, 2.38]", + "choice_d": "[6.38, 6.54]", + "answer_gt": "[6.38, 6.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 478 + }, + { + "path": "1998-29454-0016.flac", + "question": "Recognize the time when 'ain't' is mentioned by the speaker.", + "choice_a": "[0.73, 0.98]", + "choice_b": "[0.46, 0.69]", + "choice_c": "[1.33, 1.40]", + "choice_d": "[1.45, 1.60]", + "answer_gt": "[0.46, 0.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 479 + }, + { + "path": "5895-34629-0002.flac", + "question": "Note the time when 'a' is first mentioned in the audio.", + "choice_a": "[0.25, 0.57]", + "choice_b": "[1.05, 1.12]", + "choice_c": "[0.60, 0.87]", + "choice_d": "[2.08, 2.83]", + "answer_gt": "[1.05, 1.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 480 + }, + { + "path": "2609-169640-0014.flac", + "question": "Highlight when 'upon' is spoken.", + "choice_a": "[4.66, 4.91]", + "choice_b": "[0.71, 0.79]", + "choice_c": "[0.71, 0.79]", + "choice_d": "[6.51, 6.83]", + "answer_gt": "[6.51, 6.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 481 + }, + { + "path": "533-131564-0013.flac", + "question": "Establish when 'five' is first mentioned in the clip.", + "choice_a": "[0.28, 0.66]", + "choice_b": "[1.78, 2.12]", + "choice_c": "[1.26, 1.50]", + "choice_d": "[1.03, 1.26]", + "answer_gt": "[1.78, 2.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 482 + }, + { + "path": "6123-59150-0011.flac", + "question": "Highlight when 'sows' is spoken.", + "choice_a": "[1.95, 2.13]", + "choice_b": "[2.13, 2.21]", + "choice_c": "[2.21, 2.73]", + "choice_d": "[0.66, 1.03]", + "answer_gt": "[0.66, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 483 + }, + { + "path": "251-137823-0003.flac", + "question": "Identify when 'another' is heard in the audio.", + "choice_a": "[0.76, 1.20]", + "choice_b": "[2.43, 2.64]", + "choice_c": "[0.35, 0.76]", + "choice_d": "[3.01, 3.40]", + "answer_gt": "[0.35, 0.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 484 + }, + { + "path": "7021-85628-0018.flac", + "question": "Determine the point in the audio when 'get' is discussed.", + "choice_a": "[7.05, 7.27]", + "choice_b": "[6.64, 6.95]", + "choice_c": "[0.50, 0.65]", + "choice_d": "[0.82, 1.33]", + "answer_gt": "[7.05, 7.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 485 + }, + { + "path": "7641-96684-0011.flac", + "question": "Highlight the time when 'willingly' is mentioned.", + "choice_a": "[4.36, 4.62]", + "choice_b": "[0.94, 1.47]", + "choice_c": "[0.27, 0.40]", + "choice_d": "[0.74, 0.94]", + "answer_gt": "[0.94, 1.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 486 + }, + { + "path": "2033-164916-0008.flac", + "question": "Spot the exact moment 'before' appears in the audio.", + "choice_a": "[4.42, 4.78]", + "choice_b": "[5.22, 5.76]", + "choice_c": "[1.51, 2.10]", + "choice_d": "[8.05, 8.66]", + "answer_gt": "[4.42, 4.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 487 + }, + { + "path": "6070-86744-0012.flac", + "question": "When does 'fear' occur in the audio?", + "choice_a": "[1.08, 1.28]", + "choice_b": "[0.45, 0.77]", + "choice_c": "[1.47, 1.62]", + "choice_d": "[0.85, 1.08]", + "answer_gt": "[0.45, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 488 + }, + { + "path": "7127-75946-0007.flac", + "question": "Ascertain the point at which 'therefore' enters the discussion.", + "choice_a": "[3.77, 4.05]", + "choice_b": "[1.35, 1.80]", + "choice_c": "[3.63, 3.77]", + "choice_d": "[2.06, 2.25]", + "answer_gt": "[1.35, 1.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 489 + }, + { + "path": "3528-168669-0011.flac", + "question": "Highlight when 'working' is spoken.", + "choice_a": "[2.71, 3.17]", + "choice_b": "[1.58, 1.94]", + "choice_c": "[4.55, 4.80]", + "choice_d": "[3.55, 4.04]", + "answer_gt": "[1.58, 1.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 490 + }, + { + "path": "3331-159605-0022.flac", + "question": "Identify the moment 'answered' is mentioned in the audio.", + "choice_a": "[0.49, 0.64]", + "choice_b": "[1.61, 2.18]", + "choice_c": "[2.38, 2.70]", + "choice_d": "[1.18, 1.37]", + "answer_gt": "[2.38, 2.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 491 + }, + { + "path": "5543-27761-0037.flac", + "question": "Note the time when 'the' is first mentioned in the audio.", + "choice_a": "[0.73, 0.78]", + "choice_b": "[2.41, 2.74]", + "choice_c": "[0.53, 0.73]", + "choice_d": "[0.73, 0.78]", + "answer_gt": "[0.73, 0.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 492 + }, + { + "path": "7127-75947-0031.flac", + "question": "When is 'are' first heard in the recording?", + "choice_a": "[0.44, 0.59]", + "choice_b": "[0.69, 1.27]", + "choice_c": "[0.69, 1.27]", + "choice_d": "[0.59, 0.69]", + "answer_gt": "[0.59, 0.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 493 + }, + { + "path": "6123-59150-0007.flac", + "question": "Determine the point in the audio when 'dostoevsky' is repeated.", + "choice_a": "[3.40, 3.52]", + "choice_b": "[2.22, 2.51]", + "choice_c": "[1.54, 1.65]", + "choice_d": "[0.56, 1.19]", + "answer_gt": "[0.56, 1.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 494 + }, + { + "path": "5442-41169-0016.flac", + "question": "Determine the point in the audio when 'your' is discussed.", + "choice_a": "[1.79, 2.00]", + "choice_b": "[1.66, 1.79]", + "choice_c": "[2.00, 2.35]", + "choice_d": "[1.52, 1.66]", + "answer_gt": "[1.79, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 495 + }, + { + "path": "3575-170457-0036.flac", + "question": "Find out when the speaker refers to 'of' during the audio.", + "choice_a": "[5.57, 15.7]", + "choice_b": "[8.08, 8.23]", + "choice_c": "[3.04, 3.11]", + "choice_d": "[1.03, 1.20]", + "answer_gt": "[3.04, 3.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 496 + }, + { + "path": "777-126732-0023.flac", + "question": "When does 'raised' occur in the audio?", + "choice_a": "[3.02, 3.33]", + "choice_b": "[2.00, 2.10]", + "choice_c": "[4.56, 4.77]", + "choice_d": "[2.10, 2.26]", + "answer_gt": "[3.02, 3.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 497 + }, + { + "path": "5849-50873-0012.flac", + "question": "Pinpoint the exact moment 'where' appears.", + "choice_a": "[2.95, 3.33]", + "choice_b": "[2.78, 2.95]", + "choice_c": "[3.33, 3.50]", + "choice_d": "[2.16, 2.66]", + "answer_gt": "[3.33, 3.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 498 + }, + { + "path": "3081-166546-0051.flac", + "question": "Determine the time when 'the' happens within the clip.", + "choice_a": "[9.87, 10.2]", + "choice_b": "[5.07, 5.54]", + "choice_c": "[1.65, 2.13]", + "choice_d": "[7.34, 7.47]", + "answer_gt": "[7.34, 7.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 499 + }, + { + "path": "8842-302203-0005.flac", + "question": "Identify the moment 'bewilderment' is mentioned in the audio.", + "choice_a": "[1.10, 1.13]", + "choice_b": "[1.13, 1.67]", + "choice_c": "[5.38, 5.93]", + "choice_d": "[3.05, 3.28]", + "answer_gt": "[1.13, 1.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 500 + }, + { + "path": "1998-29455-0017.flac", + "question": "Identify the portion of the clip where 'and' is most prominent.", + "choice_a": "[6.70, 6.91]", + "choice_b": "[0.55, 0.61]", + "choice_c": "[4.05, 4.24]", + "choice_d": "[1.97, 2.16]", + "answer_gt": "[4.05, 4.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 501 + }, + { + "path": "1688-142285-0031.flac", + "question": "At what interval does 'stole' appear in the audio?", + "choice_a": "[5.31, 15.6]", + "choice_b": "[7.52, 8.12]", + "choice_c": "[4.44, 4.60]", + "choice_d": "[2.68, 2.93]", + "answer_gt": "[7.52, 8.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 502 + }, + { + "path": "3853-163249-0007.flac", + "question": "Pinpoint the exact moment 'kind' appears.", + "choice_a": "[4.29, 4.59]", + "choice_b": "[1.53, 1.73]", + "choice_c": "[1.73, 1.79]", + "choice_d": "[4.18, 4.29]", + "answer_gt": "[4.29, 4.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 503 + }, + { + "path": "4323-55228-0023.flac", + "question": "Recognize the time when 'wrote' is most clearly articulated.", + "choice_a": "[1.18, 1.34]", + "choice_b": "[0.37, 0.66]", + "choice_c": "[1.34, 1.40]", + "choice_d": "[1.34, 1.40]", + "answer_gt": "[1.18, 1.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 504 + }, + { + "path": "8288-274150-0007.flac", + "question": "Detect the point at which 'to' is spoken.", + "choice_a": "[6.15, 6.67]", + "choice_b": "[4.32, 4.49]", + "choice_c": "[5.55, 6.01]", + "choice_d": "[6.01, 6.15]", + "answer_gt": "[6.01, 6.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 505 + }, + { + "path": "1462-170142-0034.flac", + "question": "Listen for the precise interval when 'clean' is dominant.", + "choice_a": "[0.42, 0.62]", + "choice_b": "[0.99, 1.20]", + "choice_c": "[0.62, 0.88]", + "choice_d": "[1.25, 1.63]", + "answer_gt": "[1.25, 1.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 506 + }, + { + "path": "7975-280084-0016.flac", + "question": "Establish when 'almost' is first mentioned in the clip.", + "choice_a": "[5.72, 5.86]", + "choice_b": "[0.77, 1.29]", + "choice_c": "[1.45, 1.60]", + "choice_d": "[3.79, 4.20]", + "answer_gt": "[3.79, 4.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 507 + }, + { + "path": "1651-136854-0030.flac", + "question": "Find the moment 'in' is clearly audible in the audio.", + "choice_a": "[8.60, 8.71]", + "choice_b": "[4.79, 4.99]", + "choice_c": "[0.00, 10.1]", + "choice_d": "[0.14, 10.8]", + "answer_gt": "[0.00, 10.1]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 508 + }, + { + "path": "1188-133604-0019.flac", + "question": "What is the timestamp for when 'not' occurs?", + "choice_a": "[0.63, 10.8]", + "choice_b": "[8.76, 8.83]", + "choice_c": "[6.97, 7.09]", + "choice_d": "[0.47, 0.81]", + "answer_gt": "[0.63, 10.8]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 509 + }, + { + "path": "6295-64301-0001.flac", + "question": "At what point in the audio is 'holiday' clearly audible?", + "choice_a": "[0.60, 0.80]", + "choice_b": "[3.01, 3.18]", + "choice_c": "[0.53, 0.60]", + "choice_d": "[1.70, 2.35]", + "answer_gt": "[1.70, 2.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 510 + }, + { + "path": "4992-41806-0015.flac", + "question": "Determine the point in the audio when 'but' is discussed.", + "choice_a": "[6.37, 6.77]", + "choice_b": "[9.07, 9.20]", + "choice_c": "[7.48, 7.67]", + "choice_d": "[8.15, 8.64]", + "answer_gt": "[9.07, 9.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 511 + }, + { + "path": "8463-294825-0015.flac", + "question": "Find out the time when 'of' is immediately occured.", + "choice_a": "[1.97, 2.23]", + "choice_b": "[2.23, 2.33]", + "choice_c": "[1.44, 1.66]", + "choice_d": "[1.07, 1.44]", + "answer_gt": "[2.23, 2.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 512 + }, + { + "path": "5543-27761-0067.flac", + "question": "Recognize the time when 'aleksandrovna' is most clearly articulated.", + "choice_a": "[3.93, 4.44]", + "choice_b": "[5.29, 5.36]", + "choice_c": "[4.44, 4.57]", + "choice_d": "[1.13, 2.22]", + "answer_gt": "[1.13, 2.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 513 + }, + { + "path": "8254-115543-0001.flac", + "question": "In the audio track, when is 'was' first noticed?", + "choice_a": "[1.11, 1.43]", + "choice_b": "[0.71, 0.85]", + "choice_c": "[0.85, 1.11]", + "choice_d": "[5.47, 6.00]", + "answer_gt": "[0.71, 0.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 514 + }, + { + "path": "1995-1837-0002.flac", + "question": "Detect the point at which 'the' is spoken.", + "choice_a": "[0.62, 0.73]", + "choice_b": "[0.28, 0.62]", + "choice_c": "[1.63, 2.09]", + "choice_d": "[0.73, 1.30]", + "answer_gt": "[0.62, 0.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 515 + }, + { + "path": "6313-76958-0020.flac", + "question": "In the audio track, when is 'asked' first noticed?", + "choice_a": "[0.82, 1.09]", + "choice_b": "[0.58, 0.82]", + "choice_c": "[1.17, 1.60]", + "choice_d": "[1.60, 1.66]", + "answer_gt": "[1.17, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 516 + }, + { + "path": "1688-142285-0045.flac", + "question": "Recognize the segment where 'her' is spoken by the speaker.", + "choice_a": "[3.18, 3.51]", + "choice_b": "[0.99, 1.56]", + "choice_c": "[2.01, 2.07]", + "choice_d": "[4.56, 4.74]", + "answer_gt": "[2.01, 2.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 517 + }, + { + "path": "1585-157660-0007.flac", + "question": "When is 'sight' first mentioned in the clip?", + "choice_a": "[0.67, 0.94]", + "choice_b": "[4.40, 4.83]", + "choice_c": "[6.25, 6.69]", + "choice_d": "[0.94, 1.13]", + "answer_gt": "[0.94, 1.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 518 + }, + { + "path": "4515-11057-0089.flac", + "question": "Recognize the time when 'misery' is most clearly articulated.", + "choice_a": "[2.08, 2.33]", + "choice_b": "[1.95, 2.08]", + "choice_c": "[1.47, 1.77]", + "choice_d": "[1.00, 1.47]", + "answer_gt": "[1.00, 1.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 519 + }, + { + "path": "4294-35475-0021.flac", + "question": "Listen for the precise interval when 'rot' is dominant.", + "choice_a": "[4.05, 4.22]", + "choice_b": "[7.74, 7.84]", + "choice_c": "[3.86, 4.05]", + "choice_d": "[1.66, 1.99]", + "answer_gt": "[1.66, 1.99]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 520 + }, + { + "path": "5849-50964-0000.flac", + "question": "Discover when 'know' is introduced in the conversation.", + "choice_a": "[0.74, 0.83]", + "choice_b": "[3.19, 3.31]", + "choice_c": "[3.82, 4.23]", + "choice_d": "[1.02, 1.11]", + "answer_gt": "[0.74, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 521 + }, + { + "path": "4507-16021-0057.flac", + "question": "Find out when the speaker refers to 'people' during the audio.", + "choice_a": "[2.47, 3.25]", + "choice_b": "[1.30, 1.45]", + "choice_c": "[0.78, 1.30]", + "choice_d": "[0.46, 0.78]", + "answer_gt": "[0.46, 0.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 522 + }, + { + "path": "2428-83705-0043.flac", + "question": "Discover when 'always' is introduced in the conversation.", + "choice_a": "[3.46, 3.86]", + "choice_b": "[1.75, 2.12]", + "choice_c": "[1.47, 1.58]", + "choice_d": "[0.28, 0.73]", + "answer_gt": "[1.75, 2.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 523 + }, + { + "path": "1630-96099-0014.flac", + "question": "Determine the point in the audio when 'why' is discussed.", + "choice_a": "[0.93, 1.14]", + "choice_b": "[1.14, 1.30]", + "choice_c": "[0.20, 0.44]", + "choice_d": "[1.30, 1.37]", + "answer_gt": "[0.20, 0.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 524 + }, + { + "path": "2803-161169-0003.flac", + "question": "When does the audio feature the sound of 'the'?", + "choice_a": "[6.62, 6.91]", + "choice_b": "[5.51, 6.12]", + "choice_c": "[0.43, 0.60]", + "choice_d": "[6.91, 6.98]", + "answer_gt": "[6.91, 6.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 525 + }, + { + "path": "2300-131720-0024.flac", + "question": "Identify the portion of the clip where 'it' is most prominent.", + "choice_a": "[0.52, 0.59]", + "choice_b": "[3.83, 4.00]", + "choice_c": "[2.89, 3.20]", + "choice_d": "[1.92, 2.02]", + "answer_gt": "[1.92, 2.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 526 + }, + { + "path": "5683-32879-0002.flac", + "question": "When does 'small' occur in the audio?", + "choice_a": "[6.43, 6.74]", + "choice_b": "[2.65, 3.13]", + "choice_c": "[4.98, 5.36]", + "choice_d": "[3.24, 3.67]", + "answer_gt": "[4.98, 5.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 527 + }, + { + "path": "8131-117016-0045.flac", + "question": "Identify the moment 'to' is mentioned in the audio.", + "choice_a": "[1.63, 1.80]", + "choice_b": "[0.57, 0.68]", + "choice_c": "[1.08, 1.19]", + "choice_d": "[1.19, 1.30]", + "answer_gt": "[1.08, 1.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 528 + }, + { + "path": "3080-5040-0018.flac", + "question": "Pinpoint the exact moment 'for' is repeated in the audio.", + "choice_a": "[0.53, 1.06]", + "choice_b": "[1.21, 1.44]", + "choice_c": "[1.44, 1.55]", + "choice_d": "[1.44, 1.55]", + "answer_gt": "[1.21, 1.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 529 + }, + { + "path": "1188-133604-0011.flac", + "question": "Locate the segment where 'dawn' appears most frequently.", + "choice_a": "[9.43, 9.82]", + "choice_b": "[0.54, 0.62]", + "choice_c": "[9.89, 10.5]", + "choice_d": "[3.79, 13.9]", + "answer_gt": "[9.43, 9.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 530 + }, + { + "path": "61-70968-0031.flac", + "question": "When is 'knave' first heard in the recording?", + "choice_a": "[1.13, 1.82]", + "choice_b": "[0.21, 0.94]", + "choice_c": "[1.86, 2.14]", + "choice_d": "[1.86, 2.14]", + "answer_gt": "[1.13, 1.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 531 + }, + { + "path": "6241-61943-0023.flac", + "question": "Highlight when 'it' is spoken.", + "choice_a": "[0.48, 0.75]", + "choice_b": "[0.96, 1.45]", + "choice_c": "[1.69, 1.79]", + "choice_d": "[0.38, 0.48]", + "answer_gt": "[1.69, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 532 + }, + { + "path": "7975-280057-0020.flac", + "question": "Listen to this audio clip: when does 'i' occur?", + "choice_a": "[0.49, 0.66]", + "choice_b": "[1.96, 2.09]", + "choice_c": "[0.44, 0.49]", + "choice_d": "[0.66, 0.94]", + "answer_gt": "[0.44, 0.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 533 + }, + { + "path": "6345-64257-0014.flac", + "question": "Identify the time segment when 'he' is mentioned in the audio.", + "choice_a": "[0.18, 0.42]", + "choice_b": "[0.88, 0.95]", + "choice_c": "[1.49, 1.63]", + "choice_d": "[0.71, 0.88]", + "answer_gt": "[0.18, 0.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 534 + }, + { + "path": "5694-64029-0024.flac", + "question": "Find the duration when 'willing' occurs consecutively in the audio.", + "choice_a": "[2.28, 2.86]", + "choice_b": "[4.14, 4.24]", + "choice_c": "[1.37, 1.43]", + "choice_d": "[6.86, 7.16]", + "answer_gt": "[6.86, 7.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 535 + }, + { + "path": "3081-166546-0000.flac", + "question": "Note the time when 'longer' is first mentioned in the audio.", + "choice_a": "[3.68, 3.95]", + "choice_b": "[3.95, 4.34]", + "choice_c": "[5.83, 6.30]", + "choice_d": "[2.22, 2.29]", + "answer_gt": "[5.83, 6.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 536 + }, + { + "path": "2803-161169-0010.flac", + "question": "Locate the segment where 'in' appears most frequently.", + "choice_a": "[5.39, 5.50]", + "choice_b": "[9.11, 9.33]", + "choice_c": "[2.52, 12.9]", + "choice_d": "[7.77, 7.95]", + "answer_gt": "[5.39, 5.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 537 + }, + { + "path": "6432-63723-0046.flac", + "question": "Pinpoint the passage where 'of' is heard in the background.", + "choice_a": "[4.17, 4.28]", + "choice_b": "[0.86, 0.98]", + "choice_c": "[2.99, 3.48]", + "choice_d": "[2.74, 2.99]", + "answer_gt": "[0.86, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 538 + }, + { + "path": "7105-2340-0026.flac", + "question": "Recognize the time when 'you' is most clearly articulated.", + "choice_a": "[0.51, 0.74]", + "choice_b": "[4.41, 4.48]", + "choice_c": "[1.16, 11.5]", + "choice_d": "[1.66, 1.76]", + "answer_gt": "[1.66, 1.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 539 + }, + { + "path": "8188-269290-0045.flac", + "question": "Establish when 'no' is first mentioned in the clip.", + "choice_a": "[3.79, 4.02]", + "choice_b": "[0.82, 1.20]", + "choice_c": "[2.70, 2.85]", + "choice_d": "[1.46, 1.60]", + "answer_gt": "[1.46, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 540 + }, + { + "path": "6467-94831-0020.flac", + "question": "Pinpoint the exact moment 'the' appears.", + "choice_a": "[1.34, 1.50]", + "choice_b": "[3.61, 3.66]", + "choice_c": "[1.16, 1.34]", + "choice_d": "[0.37, 0.60]", + "answer_gt": "[3.61, 3.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 541 + }, + { + "path": "672-122797-0038.flac", + "question": "What is the timestamp for when 'thought' occurs?", + "choice_a": "[4.46, 4.53]", + "choice_b": "[2.63, 3.07]", + "choice_c": "[0.00, 0.34]", + "choice_d": "[5.57, 6.02]", + "answer_gt": "[0.00, 0.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 542 + }, + { + "path": "1998-15444-0024.flac", + "question": "At what time interval does 'glass' occur in the audio clip?", + "choice_a": "[2.27, 2.99]", + "choice_b": "[1.80, 1.96]", + "choice_c": "[4.56, 4.96]", + "choice_d": "[5.43, 5.73]", + "answer_gt": "[4.56, 4.96]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 543 + }, + { + "path": "8461-281231-0038.flac", + "question": "What is the timestamp for the initial occurrence of 'seats'?", + "choice_a": "[2.64, 3.23]", + "choice_b": "[9.00, 9.13]", + "choice_c": "[4.07, 4.46]", + "choice_d": "[4.86, 4.92]", + "answer_gt": "[4.07, 4.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 544 + }, + { + "path": "2277-149874-0011.flac", + "question": "Establish when 'of' is first mentioned in the clip.", + "choice_a": "[0.33, 0.61]", + "choice_b": "[1.21, 1.31]", + "choice_c": "[0.70, 0.83]", + "choice_d": "[0.83, 1.21]", + "answer_gt": "[1.21, 1.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 545 + }, + { + "path": "6841-88294-0053.flac", + "question": "Establish when 'and' is first mentioned in the clip.", + "choice_a": "[0.72, 0.86]", + "choice_b": "[1.96, 2.24]", + "choice_c": "[3.40, 3.69]", + "choice_d": "[1.43, 1.54]", + "answer_gt": "[1.43, 1.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 546 + }, + { + "path": "4198-12259-0012.flac", + "question": "What is the timestamp for the initial occurrence of 'and'?", + "choice_a": "[1.09, 1.24]", + "choice_b": "[1.93, 2.26]", + "choice_c": "[1.46, 1.93]", + "choice_d": "[1.24, 1.46]", + "answer_gt": "[1.09, 1.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 547 + }, + { + "path": "4572-64670-0008.flac", + "question": "Determine the point in the audio when 'great' is repeated.", + "choice_a": "[8.34, 8.52]", + "choice_b": "[7.44, 7.88]", + "choice_c": "[0.86, 11.1]", + "choice_d": "[1.53, 1.92]", + "answer_gt": "[0.86, 11.1]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 548 + }, + { + "path": "2414-159411-0011.flac", + "question": "What is the timestamp for when 'a' occurs?", + "choice_a": "[0.55, 0.60]", + "choice_b": "[1.78, 1.92]", + "choice_c": "[3.87, 4.00]", + "choice_d": "[0.97, 1.60]", + "answer_gt": "[0.55, 0.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 549 + }, + { + "path": "5142-33396-0009.flac", + "question": "At what time interval does 'glare' occur in the audio clip?", + "choice_a": "[0.14, 0.55]", + "choice_b": "[1.74, 2.09]", + "choice_c": "[2.20, 2.40]", + "choice_d": "[0.55, 0.93]", + "answer_gt": "[1.74, 2.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 550 + }, + { + "path": "4350-10919-0018.flac", + "question": "Note the time when 'ah' is first mentioned in the audio.", + "choice_a": "[2.74, 3.10]", + "choice_b": "[0.45, 0.80]", + "choice_c": "[1.97, 2.10]", + "choice_d": "[1.27, 1.43]", + "answer_gt": "[0.45, 0.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 551 + }, + { + "path": "1089-134686-0022.flac", + "question": "Identify the portion of the clip where 'of' is most prominent.", + "choice_a": "[1.68, 1.77]", + "choice_b": "[3.96, 4.10]", + "choice_c": "[4.55, 4.87]", + "choice_d": "[9.51, 9.75]", + "answer_gt": "[3.96, 4.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 552 + }, + { + "path": "3997-180294-0029.flac", + "question": "Identify the portion of the clip where 'that' is most prominent.", + "choice_a": "[1.03, 1.17]", + "choice_b": "[2.83, 3.03]", + "choice_c": "[5.04, 5.37]", + "choice_d": "[2.66, 2.83]", + "answer_gt": "[2.83, 3.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 553 + }, + { + "path": "6455-67803-0019.flac", + "question": "Pinpoint the passage where 'a' is heard in the background.", + "choice_a": "[0.48, 0.87]", + "choice_b": "[5.08, 5.16]", + "choice_c": "[1.32, 1.38]", + "choice_d": "[5.27, 5.33]", + "answer_gt": "[5.27, 5.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 554 + }, + { + "path": "1650-157641-0012.flac", + "question": "Recognize the time when 'at' is most clearly articulated.", + "choice_a": "[1.66, 1.78]", + "choice_b": "[0.25, 0.84]", + "choice_c": "[1.39, 1.66]", + "choice_d": "[1.39, 1.66]", + "answer_gt": "[1.66, 1.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 555 + }, + { + "path": "1221-135767-0010.flac", + "question": "Pinpoint the exact moment 'the' is repeated in the audio.", + "choice_a": "[5.26, 5.69]", + "choice_b": "[5.21, 5.26]", + "choice_c": "[4.43, 4.86]", + "choice_d": "[4.86, 5.21]", + "answer_gt": "[5.21, 5.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 556 + }, + { + "path": "3997-182399-0016.flac", + "question": "Pinpoint the exact moment 'yes' appears.", + "choice_a": "[0.27, 0.54]", + "choice_b": "[1.78, 2.00]", + "choice_c": "[1.28, 1.78]", + "choice_d": "[0.54, 1.11]", + "answer_gt": "[0.27, 0.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 557 + }, + { + "path": "2094-142345-0033.flac", + "question": "At what interval does 'and' appear in the audio?", + "choice_a": "[6.68, 6.81]", + "choice_b": "[2.13, 2.23]", + "choice_c": "[2.23, 2.34]", + "choice_d": "[9.43, 9.55]", + "answer_gt": "[2.13, 2.23]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 558 + }, + { + "path": "3576-138058-0024.flac", + "question": "Spot the exact moment 'she' appears in the audio.", + "choice_a": "[8.47, 8.61]", + "choice_b": "[2.50, 2.69]", + "choice_c": "[3.99, 4.07]", + "choice_d": "[1.50, 21.5]", + "answer_gt": "[2.50, 2.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 559 + }, + { + "path": "1221-135766-0004.flac", + "question": "Highlight the time when 'this' is mentioned.", + "choice_a": "[0.54, 0.75]", + "choice_b": "[3.03, 3.29]", + "choice_c": "[6.51, 6.74]", + "choice_d": "[4.99, 5.04]", + "answer_gt": "[0.54, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 560 + }, + { + "path": "1650-157641-0004.flac", + "question": "Locate the part of the audio where 'in' is expressed.", + "choice_a": "[0.59, 0.93]", + "choice_b": "[0.37, 0.59]", + "choice_c": "[1.50, 1.59]", + "choice_d": "[1.89, 2.02]", + "answer_gt": "[1.89, 2.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 561 + }, + { + "path": "908-31957-0001.flac", + "question": "At what time interval does 'inward' occur in the audio clip?", + "choice_a": "[0.08, 10.5]", + "choice_b": "[8.12, 8.92]", + "choice_c": "[1.27, 1.78]", + "choice_d": "[2.70, 2.91]", + "answer_gt": "[0.08, 10.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 562 + }, + { + "path": "2428-83705-0008.flac", + "question": "When does the audio feature the sound of 'i'?", + "choice_a": "[0.45, 0.57]", + "choice_b": "[1.84, 2.66]", + "choice_c": "[1.84, 2.66]", + "choice_d": "[1.84, 2.66]", + "answer_gt": "[0.45, 0.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 563 + }, + { + "path": "6930-75918-0014.flac", + "question": "Listen for when 'who' is emphasized by the narrator.", + "choice_a": "[0.46, 0.92]", + "choice_b": "[2.54, 2.89]", + "choice_c": "[7.07, 7.22]", + "choice_d": "[9.01, 9.70]", + "answer_gt": "[7.07, 7.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 564 + }, + { + "path": "3764-168670-0053.flac", + "question": "Identify when 'is' is heard in the audio.", + "choice_a": "[1.48, 1.61]", + "choice_b": "[0.38, 0.58]", + "choice_c": "[1.23, 1.48]", + "choice_d": "[0.25, 0.38]", + "answer_gt": "[1.48, 1.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 565 + }, + { + "path": "7697-105815-0044.flac", + "question": "When is 'ask' first heard in the recording?", + "choice_a": "[3.73, 4.09]", + "choice_b": "[3.35, 3.55]", + "choice_c": "[4.63, 4.92]", + "choice_d": "[3.55, 3.73]", + "answer_gt": "[4.63, 4.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 566 + }, + { + "path": "251-136532-0020.flac", + "question": "When does the speaker mention 'public'?", + "choice_a": "[0.61, 0.70]", + "choice_b": "[1.14, 1.73]", + "choice_c": "[0.30, 0.52]", + "choice_d": "[0.78, 1.14]", + "answer_gt": "[0.78, 1.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 567 + }, + { + "path": "260-123286-0016.flac", + "question": "Determine the point in the audio when 'agitated' is repeated.", + "choice_a": "[4.24, 4.60]", + "choice_b": "[4.60, 4.95]", + "choice_c": "[1.23, 1.79]", + "choice_d": "[2.99, 3.65]", + "answer_gt": "[1.23, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 568 + }, + { + "path": "1320-122617-0040.flac", + "question": "Pinpoint the exact moment 'next' is repeated in the audio.", + "choice_a": "[2.33, 2.82]", + "choice_b": "[6.58, 6.96]", + "choice_c": "[2.25, 2.33]", + "choice_d": "[5.67, 5.90]", + "answer_gt": "[2.33, 2.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 569 + }, + { + "path": "8842-302201-0013.flac", + "question": "At what point in the audio is 'is' clearly audible?", + "choice_a": "[0.27, 0.38]", + "choice_b": "[0.38, 0.43]", + "choice_c": "[4.25, 4.36]", + "choice_d": "[1.41, 1.53]", + "answer_gt": "[1.41, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 570 + }, + { + "path": "7641-96670-0024.flac", + "question": "Determine the precise moment 'jason' in the audio's pace.", + "choice_a": "[0.94, 1.46]", + "choice_b": "[1.96, 2.65]", + "choice_c": "[3.46, 3.72]", + "choice_d": "[3.72, 3.78]", + "answer_gt": "[0.94, 1.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 571 + }, + { + "path": "3081-166546-0033.flac", + "question": "When is 'absolutely' first heard in the recording?", + "choice_a": "[0.32, 0.56]", + "choice_b": "[1.05, 1.28]", + "choice_c": "[0.56, 1.05]", + "choice_d": "[1.34, 2.13]", + "answer_gt": "[1.34, 2.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 572 + }, + { + "path": "2086-149220-0049.flac", + "question": "At what interval does 'as' appear in the audio?", + "choice_a": "[3.30, 3.43]", + "choice_b": "[3.89, 4.01]", + "choice_c": "[3.19, 3.30]", + "choice_d": "[3.43, 3.53]", + "answer_gt": "[3.30, 3.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 573 + }, + { + "path": "4350-9170-0028.flac", + "question": "Ascertain the point at which 'in' enters the discussion.", + "choice_a": "[8.71, 9.04]", + "choice_b": "[9.30, 9.70]", + "choice_c": "[4.14, 4.22]", + "choice_d": "[5.21, 5.31]", + "answer_gt": "[4.14, 4.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 574 + }, + { + "path": "533-131556-0014.flac", + "question": "Discover when 'her' is introduced in the conversation.", + "choice_a": "[5.84, 6.22]", + "choice_b": "[8.00, 8.43]", + "choice_c": "[0.80, 0.88]", + "choice_d": "[8.62, 8.75]", + "answer_gt": "[0.80, 0.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 575 + }, + { + "path": "4153-186222-0019.flac", + "question": "Find out when the speaker refers to 'world' during the audio.", + "choice_a": "[1.09, 1.27]", + "choice_b": "[0.30, 0.44]", + "choice_c": "[2.37, 2.73]", + "choice_d": "[0.44, 0.72]", + "answer_gt": "[2.37, 2.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 576 + }, + { + "path": "1630-102884-0011.flac", + "question": "At what point in the audio is 'harmony' clearly audible?", + "choice_a": "[7.46, 7.92]", + "choice_b": "[4.15, 4.34]", + "choice_c": "[1.59, 2.76]", + "choice_d": "[6.57, 7.15]", + "answer_gt": "[6.57, 7.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 577 + }, + { + "path": "7976-110523-0006.flac", + "question": "Pinpoint the exact moment 'he' appears.", + "choice_a": "[7.01, 7.06]", + "choice_b": "[6.03, 6.11]", + "choice_c": "[1.98, 2.24]", + "choice_d": "[4.27, 4.35]", + "answer_gt": "[4.27, 4.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 578 + }, + { + "path": "3528-168669-0100.flac", + "question": "When is 'then' first mentioned in the clip?", + "choice_a": "[0.80, 0.95]", + "choice_b": "[1.59, 1.66]", + "choice_c": "[1.40, 1.59]", + "choice_d": "[1.30, 1.40]", + "answer_gt": "[1.40, 1.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 579 + }, + { + "path": "2033-164915-0012.flac", + "question": "Pinpoint the exact moment 'pages' is repeated in the audio.", + "choice_a": "[4.99, 5.10]", + "choice_b": "[5.10, 5.34]", + "choice_c": "[2.21, 2.84]", + "choice_d": "[1.75, 2.13]", + "answer_gt": "[2.21, 2.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 580 + }, + { + "path": "6455-67804-0005.flac", + "question": "Determine the precise moment 'comment' in the audio's pace.", + "choice_a": "[0.39, 0.63]", + "choice_b": "[2.34, 2.49]", + "choice_c": "[1.48, 1.84]", + "choice_d": "[4.62, 5.34]", + "answer_gt": "[1.48, 1.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 581 + }, + { + "path": "8224-274384-0000.flac", + "question": "When does 'the' occur in the audio?", + "choice_a": "[4.42, 4.70]", + "choice_b": "[6.33, 6.56]", + "choice_c": "[0.58, 1.04]", + "choice_d": "[6.56, 6.61]", + "answer_gt": "[6.56, 6.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 582 + }, + { + "path": "6841-88291-0053.flac", + "question": "Note the time interval that includes the mention of 'the'.", + "choice_a": "[1.85, 1.96]", + "choice_b": "[0.31, 0.39]", + "choice_c": "[0.39, 0.72]", + "choice_d": "[1.41, 1.52]", + "answer_gt": "[0.31, 0.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 583 + }, + { + "path": "5895-34622-0011.flac", + "question": "Recognize the segment where 'after' is spoken by the speaker.", + "choice_a": "[2.67, 2.97]", + "choice_b": "[3.75, 4.48]", + "choice_c": "[2.97, 3.37]", + "choice_d": "[4.28, 14.5]", + "answer_gt": "[4.28, 14.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 584 + }, + { + "path": "5683-32866-0022.flac", + "question": "When in the audio does 'its' coincide with a specific event?", + "choice_a": "[1.67, 2.00]", + "choice_b": "[0.91, 1.06]", + "choice_c": "[0.26, 0.42]", + "choice_d": "[1.06, 1.15]", + "answer_gt": "[0.26, 0.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 585 + }, + { + "path": "3660-172182-0039.flac", + "question": "Determine the point in the audio when 'homage' is repeated.", + "choice_a": "[0.36, 0.60]", + "choice_b": "[1.11, 1.23]", + "choice_c": "[1.23, 1.54]", + "choice_d": "[1.96, 2.13]", + "answer_gt": "[1.23, 1.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 586 + }, + { + "path": "6345-93302-0021.flac", + "question": "Listen for the occurrence of 'sorry' in the recording.", + "choice_a": "[0.90, 1.11]", + "choice_b": "[1.11, 1.40]", + "choice_c": "[0.78, 0.90]", + "choice_d": "[0.57, 0.78]", + "answer_gt": "[1.11, 1.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 587 + }, + { + "path": "2830-3980-0042.flac", + "question": "Pinpoint the passage where 'pernicious' is heard in the background.", + "choice_a": "[1.65, 2.19]", + "choice_b": "[1.36, 1.60]", + "choice_c": "[1.01, 1.36]", + "choice_d": "[0.64, 1.01]", + "answer_gt": "[1.65, 2.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 588 + }, + { + "path": "8842-302203-0001.flac", + "question": "Detect the point at which 'voice' is spoken.", + "choice_a": "[0.31, 0.44]", + "choice_b": "[4.73, 5.04]", + "choice_c": "[5.04, 5.54]", + "choice_d": "[7.57, 7.77]", + "answer_gt": "[5.04, 5.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 589 + }, + { + "path": "5105-28241-0016.flac", + "question": "Spot the exact moment 'you' appears in the audio.", + "choice_a": "[4.26, 4.97]", + "choice_b": "[5.03, 5.17]", + "choice_c": "[0.56, 0.69]", + "choice_d": "[2.02, 2.09]", + "answer_gt": "[0.56, 0.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 590 + }, + { + "path": "8288-274162-0015.flac", + "question": "When is 'he' first heard in the recording?", + "choice_a": "[0.32, 0.44]", + "choice_b": "[0.44, 0.58]", + "choice_c": "[1.90, 2.00]", + "choice_d": "[1.00, 1.32]", + "answer_gt": "[1.90, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 591 + }, + { + "path": "1462-170138-0002.flac", + "question": "Locate the segment where 'our' is faintly heard.", + "choice_a": "[1.58, 1.93]", + "choice_b": "[3.50, 3.62]", + "choice_c": "[0.72, 0.84]", + "choice_d": "[0.44, 0.72]", + "answer_gt": "[3.50, 3.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 592 + }, + { + "path": "5639-40744-0036.flac", + "question": "Listen to this audio clip; identify the time segment when 'so' is spoken.", + "choice_a": "[5.30, 5.44]", + "choice_b": "[9.34, 9.66]", + "choice_c": "[1.88, 2.01]", + "choice_d": "[4.26, 4.44]", + "answer_gt": "[9.34, 9.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 593 + }, + { + "path": "5142-33396-0019.flac", + "question": "Listen for when 'better' is emphasized by the narrator.", + "choice_a": "[1.30, 1.36]", + "choice_b": "[0.25, 0.59]", + "choice_c": "[2.24, 2.34]", + "choice_d": "[0.95, 1.30]", + "answer_gt": "[0.95, 1.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 594 + }, + { + "path": "2300-131720-0008.flac", + "question": "Highlight when 'energy' is spoken.", + "choice_a": "[3.70, 4.20]", + "choice_b": "[3.59, 3.70]", + "choice_c": "[0.88, 1.07]", + "choice_d": "[3.59, 3.70]", + "answer_gt": "[3.70, 4.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 595 + }, + { + "path": "7976-110124-0017.flac", + "question": "Ascertain the point at which 'perhaps' enters the discussion.", + "choice_a": "[1.15, 1.52]", + "choice_b": "[1.90, 2.14]", + "choice_c": "[0.49, 0.92]", + "choice_d": "[0.98, 1.12]", + "answer_gt": "[0.49, 0.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 596 + }, + { + "path": "908-157963-0019.flac", + "question": "Identify the portion of the clip where 'balmy' is most prominent.", + "choice_a": "[0.13, 0.25]", + "choice_b": "[0.89, 11.2]", + "choice_c": "[7.45, 17.8]", + "choice_d": "[7.86, 18.5]", + "answer_gt": "[0.89, 11.2]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 597 + }, + { + "path": "251-118436-0006.flac", + "question": "Listen for the occurrence of 'a' in the recording.", + "choice_a": "[1.81, 1.86]", + "choice_b": "[5.34, 5.77]", + "choice_c": "[1.14, 1.33]", + "choice_d": "[6.14, 6.34]", + "answer_gt": "[1.81, 1.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 598 + }, + { + "path": "2414-128291-0016.flac", + "question": "Identify the portion of the clip where 'than' is most prominent.", + "choice_a": "[0.34, 0.67]", + "choice_b": "[3.33, 3.51]", + "choice_c": "[0.70, 1.15]", + "choice_d": "[2.49, 2.94]", + "answer_gt": "[3.33, 3.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 599 + }, + { + "path": "8254-84205-0019.flac", + "question": "Locate the segment where 'em' appears most frequently.", + "choice_a": "[1.00, 1.22]", + "choice_b": "[1.68, 1.87]", + "choice_c": "[4.08, 4.38]", + "choice_d": "[3.03, 3.21]", + "answer_gt": "[1.00, 1.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 600 + }, + { + "path": "6345-93302-0007.flac", + "question": "Determine the precise moment 'was' in the audio's pace.", + "choice_a": "[0.41, 0.50]", + "choice_b": "[3.19, 3.25]", + "choice_c": "[1.64, 2.17]", + "choice_d": "[0.50, 0.63]", + "answer_gt": "[0.50, 0.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 601 + }, + { + "path": "6313-66125-0009.flac", + "question": "Establish when 'i' is first mentioned in the clip.", + "choice_a": "[0.33, 0.63]", + "choice_b": "[0.71, 1.56]", + "choice_c": "[1.94, 2.00]", + "choice_d": "[1.94, 2.00]", + "answer_gt": "[0.33, 0.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 602 + }, + { + "path": "3000-15664-0018.flac", + "question": "Listen to this audio clip; identify the time segment when 'slopes' is spoken.", + "choice_a": "[0.33, 0.41]", + "choice_b": "[0.33, 0.41]", + "choice_c": "[0.94, 1.53]", + "choice_d": "[3.13, 3.41]", + "answer_gt": "[0.94, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 603 + }, + { + "path": "7105-2340-0008.flac", + "question": "Locate the segment where 'from' appears most frequently.", + "choice_a": "[5.86, 6.16]", + "choice_b": "[0.34, 10.6]", + "choice_c": "[2.03, 2.49]", + "choice_d": "[1.22, 11.4]", + "answer_gt": "[1.22, 11.4]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 604 + }, + { + "path": "8461-281231-0032.flac", + "question": "Find out when the speaker refers to 'visit' during the audio.", + "choice_a": "[6.36, 7.03]", + "choice_b": "[2.27, 2.52]", + "choice_c": "[7.78, 8.16]", + "choice_d": "[1.37, 2.17]", + "answer_gt": "[7.78, 8.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 605 + }, + { + "path": "5543-27761-0000.flac", + "question": "Note the time interval that includes the mention of 'lelechka'.", + "choice_a": "[1.04, 1.08]", + "choice_b": "[0.37, 0.88]", + "choice_c": "[1.08, 1.66]", + "choice_d": "[1.08, 1.66]", + "answer_gt": "[0.37, 0.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 606 + }, + { + "path": "6930-76324-0022.flac", + "question": "Note the time interval that includes the mention of 'suddenly'.", + "choice_a": "[0.76, 0.97]", + "choice_b": "[0.28, 0.73]", + "choice_c": "[0.28, 0.73]", + "choice_d": "[0.97, 1.35]", + "answer_gt": "[0.97, 1.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 607 + }, + { + "path": "1988-24833-0015.flac", + "question": "When in the audio does 'it' coincide with a specific event?", + "choice_a": "[0.69, 1.23]", + "choice_b": "[3.29, 3.39]", + "choice_c": "[0.32, 0.45]", + "choice_d": "[1.48, 1.85]", + "answer_gt": "[3.29, 3.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 608 + }, + { + "path": "260-123288-0028.flac", + "question": "Note the time interval that includes the mention of 'suffer'.", + "choice_a": "[0.85, 1.51]", + "choice_b": "[0.33, 0.43]", + "choice_c": "[0.33, 0.43]", + "choice_d": "[0.43, 0.85]", + "answer_gt": "[0.43, 0.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 609 + }, + { + "path": "6599-38590-0000.flac", + "question": "Recognize the time when 'licentious' is most clearly articulated.", + "choice_a": "[7.77, 8.47]", + "choice_b": "[1.73, 1.97]", + "choice_c": "[1.09, 1.36]", + "choice_d": "[3.27, 3.61]", + "answer_gt": "[7.77, 8.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 610 + }, + { + "path": "7105-2340-0002.flac", + "question": "When does 'make' occur in the audio?", + "choice_a": "[4.36, 5.04]", + "choice_b": "[3.94, 4.05]", + "choice_c": "[4.29, 4.36]", + "choice_d": "[4.05, 4.29]", + "answer_gt": "[4.05, 4.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 611 + }, + { + "path": "3331-159605-0037.flac", + "question": "Identify when 'dear' is heard in the audio.", + "choice_a": "[3.46, 3.52]", + "choice_b": "[3.56, 3.85]", + "choice_c": "[4.54, 4.59]", + "choice_d": "[4.29, 4.54]", + "answer_gt": "[3.56, 3.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 612 + }, + { + "path": "2277-149896-0012.flac", + "question": "When does 'to' occur in the audio?", + "choice_a": "[0.47, 0.65]", + "choice_b": "[0.36, 0.47]", + "choice_c": "[1.02, 1.15]", + "choice_d": "[0.77, 1.02]", + "answer_gt": "[1.02, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 613 + }, + { + "path": "7601-291468-0004.flac", + "question": "Recognize the time when 'had' is most clearly articulated.", + "choice_a": "[4.72, 15.1]", + "choice_b": "[1.00, 1.70]", + "choice_c": "[5.76, 5.98]", + "choice_d": "[4.34, 14.6]", + "answer_gt": "[5.76, 5.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 614 + }, + { + "path": "3005-163391-0008.flac", + "question": "Identify the time segment when 'mile' is mentioned in the audio.", + "choice_a": "[5.09, 5.23]", + "choice_b": "[8.50, 8.97]", + "choice_c": "[9.18, 9.26]", + "choice_d": "[6.04, 6.24]", + "answer_gt": "[8.50, 8.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 615 + }, + { + "path": "5142-36377-0004.flac", + "question": "Pinpoint the exact moment 'place' is repeated in the audio.", + "choice_a": "[4.25, 4.36]", + "choice_b": "[3.26, 3.72]", + "choice_c": "[3.72, 4.22]", + "choice_d": "[1.30, 1.36]", + "answer_gt": "[3.72, 4.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 616 + }, + { + "path": "2902-9008-0005.flac", + "question": "Find the duration when 'herd' occurs consecutively in the audio.", + "choice_a": "[0.36, 0.58]", + "choice_b": "[2.11, 2.46]", + "choice_c": "[1.27, 1.53]", + "choice_d": "[0.58, 0.70]", + "answer_gt": "[2.11, 2.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 617 + }, + { + "path": "8455-210777-0011.flac", + "question": "Determine the point in the audio when 'subjects' is repeated.", + "choice_a": "[2.03, 2.68]", + "choice_b": "[3.10, 3.17]", + "choice_c": "[4.12, 4.68]", + "choice_d": "[4.71, 4.89]", + "answer_gt": "[4.12, 4.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 618 + }, + { + "path": "1255-138279-0022.flac", + "question": "Identify the moment 'said' is mentioned in the audio.", + "choice_a": "[1.16, 1.35]", + "choice_b": "[0.52, 0.59]", + "choice_c": "[1.44, 1.68]", + "choice_d": "[0.59, 0.79]", + "answer_gt": "[0.59, 0.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 619 + }, + { + "path": "8555-284449-0011.flac", + "question": "When does the speaker mention 'animal'?", + "choice_a": "[7.17, 7.62]", + "choice_b": "[4.08, 4.22]", + "choice_c": "[1.57, 11.7]", + "choice_d": "[9.69, 9.73]", + "answer_gt": "[7.17, 7.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 620 + }, + { + "path": "908-31957-0019.flac", + "question": "Find the duration when 'and' occurs consecutively in the audio.", + "choice_a": "[6.48, 6.87]", + "choice_b": "[1.89, 2.09]", + "choice_c": "[3.69, 3.79]", + "choice_d": "[2.65, 2.76]", + "answer_gt": "[2.65, 2.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 621 + }, + { + "path": "1988-24833-0007.flac", + "question": "Recognize the time when 'for' is mentioned by the speaker.", + "choice_a": "[0.16, 0.30]", + "choice_b": "[1.21, 1.36]", + "choice_c": "[2.29, 2.33]", + "choice_d": "[3.80, 4.15]", + "answer_gt": "[1.21, 1.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 622 + }, + { + "path": "1988-24833-0001.flac", + "question": "What is the duration of 'fun' being mentioned in the conversation?", + "choice_a": "[0.31, 0.74]", + "choice_b": "[6.69, 6.91]", + "choice_c": "[5.13, 5.44]", + "choice_d": "[0.74, 0.93]", + "answer_gt": "[5.13, 5.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 623 + }, + { + "path": "5536-43358-0008.flac", + "question": "Recognize the segment where 'his' is spoken by the speaker.", + "choice_a": "[1.24, 11.8]", + "choice_b": "[2.76, 2.83]", + "choice_c": "[3.20, 3.35]", + "choice_d": "[2.83, 3.04]", + "answer_gt": "[3.20, 3.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 624 + }, + { + "path": "1462-170142-0015.flac", + "question": "Listen for the occurrence of 'unbearable' in the recording.", + "choice_a": "[0.15, 0.28]", + "choice_b": "[2.00, 2.23]", + "choice_c": "[0.31, 1.02]", + "choice_d": "[1.27, 1.40]", + "answer_gt": "[0.31, 1.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 625 + }, + { + "path": "1998-29455-0034.flac", + "question": "At what point in the audio is 'see' clearly audible?", + "choice_a": "[0.85, 0.96]", + "choice_b": "[1.16, 1.43]", + "choice_c": "[0.49, 0.85]", + "choice_d": "[0.49, 0.85]", + "answer_gt": "[1.16, 1.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 626 + }, + { + "path": "8555-284447-0005.flac", + "question": "Listen for the occurrence of 'around' in the recording.", + "choice_a": "[0.35, 0.43]", + "choice_b": "[3.01, 3.40]", + "choice_c": "[5.11, 5.71]", + "choice_d": "[0.43, 0.65]", + "answer_gt": "[3.01, 3.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 627 + }, + { + "path": "1650-167613-0003.flac", + "question": "Identify the portion of the clip where 'understand' is most prominent.", + "choice_a": "[2.65, 2.91]", + "choice_b": "[2.48, 2.65]", + "choice_c": "[0.89, 1.53]", + "choice_d": "[0.34, 0.74]", + "answer_gt": "[0.89, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 628 + }, + { + "path": "8463-294825-0005.flac", + "question": "Ascertain the exact time 'other' is mentioned in the audio.", + "choice_a": "[5.34, 5.69]", + "choice_b": "[3.62, 4.34]", + "choice_c": "[0.46, 0.68]", + "choice_d": "[2.22, 2.51]", + "answer_gt": "[0.46, 0.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 629 + }, + { + "path": "2094-142345-0011.flac", + "question": "Recognize the time when 'of' is most clearly articulated.", + "choice_a": "[1.16, 11.2]", + "choice_b": "[9.20, 9.48]", + "choice_c": "[3.14, 3.32]", + "choice_d": "[7.20, 7.28]", + "answer_gt": "[1.16, 11.2]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 630 + }, + { + "path": "8555-284447-0015.flac", + "question": "Locate the segment where 'of' appears most frequently.", + "choice_a": "[7.52, 7.65]", + "choice_b": "[1.37, 1.44]", + "choice_c": "[3.11, 3.17]", + "choice_d": "[3.55, 3.66]", + "answer_gt": "[3.55, 3.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 631 + }, + { + "path": "8463-287645-0003.flac", + "question": "What is the timestamp for when 'boy' occurs?", + "choice_a": "[1.70, 2.18]", + "choice_b": "[4.57, 4.86]", + "choice_c": "[2.27, 2.55]", + "choice_d": "[2.21, 2.27]", + "answer_gt": "[2.27, 2.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 632 + }, + { + "path": "5895-34615-0016.flac", + "question": "Ascertain the exact time 'and' is mentioned in the audio.", + "choice_a": "[0.41, 10.7]", + "choice_b": "[2.50, 2.65]", + "choice_c": "[6.94, 7.08]", + "choice_d": "[2.07, 12.3]", + "answer_gt": "[6.94, 7.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 633 + }, + { + "path": "2830-3980-0036.flac", + "question": "Pinpoint the exact moment 'there' appears.", + "choice_a": "[2.22, 2.36]", + "choice_b": "[2.36, 2.50]", + "choice_c": "[3.95, 4.13]", + "choice_d": "[4.13, 4.83]", + "answer_gt": "[2.22, 2.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 634 + }, + { + "path": "3538-163622-0024.flac", + "question": "Discover when 'have' is introduced in the conversation.", + "choice_a": "[0.41, 0.51]", + "choice_b": "[0.67, 0.82]", + "choice_c": "[0.51, 0.67]", + "choice_d": "[0.82, 1.17]", + "answer_gt": "[0.41, 0.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 635 + }, + { + "path": "3005-163390-0029.flac", + "question": "Discover when 'make' is introduced in the conversation.", + "choice_a": "[0.41, 0.58]", + "choice_b": "[1.16, 1.57]", + "choice_c": "[1.57, 1.69]", + "choice_d": "[0.58, 0.96]", + "answer_gt": "[1.16, 1.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 636 + }, + { + "path": "700-122866-0017.flac", + "question": "Determine the precise moment 'white' in the audio's pace.", + "choice_a": "[2.18, 2.28]", + "choice_b": "[0.77, 1.08]", + "choice_c": "[0.35, 0.44]", + "choice_d": "[3.20, 3.58]", + "answer_gt": "[0.77, 1.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 637 + }, + { + "path": "1988-148538-0013.flac", + "question": "Highlight the time when 'they' is mentioned.", + "choice_a": "[4.65, 4.81]", + "choice_b": "[1.41, 1.51]", + "choice_c": "[0.35, 0.43]", + "choice_d": "[8.72, 9.07]", + "answer_gt": "[0.35, 0.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 638 + }, + { + "path": "6467-56885-0004.flac", + "question": "Listen for the occurrence of 'his' in the recording.", + "choice_a": "[2.48, 2.63]", + "choice_b": "[4.75, 4.80]", + "choice_c": "[3.43, 3.57]", + "choice_d": "[1.93, 2.15]", + "answer_gt": "[2.48, 2.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 639 + }, + { + "path": "7697-245712-0009.flac", + "question": "Determine the precise moment 'would' in the audio's pace.", + "choice_a": "[4.32, 4.80]", + "choice_b": "[2.05, 2.25]", + "choice_c": "[1.12, 1.29]", + "choice_d": "[3.39, 3.80]", + "answer_gt": "[1.12, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 640 + }, + { + "path": "84-121123-0001.flac", + "question": "Locate the segment where 'an' appears most frequently.", + "choice_a": "[0.71, 0.85]", + "choice_b": "[1.55, 2.06]", + "choice_c": "[1.10, 1.44]", + "choice_d": "[2.78, 2.86]", + "answer_gt": "[2.78, 2.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 641 + }, + { + "path": "7176-88083-0000.flac", + "question": "In the audio track, when is 'him' first noticed?", + "choice_a": "[0.59, 0.97]", + "choice_b": "[0.25, 0.59]", + "choice_c": "[1.10, 1.24]", + "choice_d": "[0.97, 1.10]", + "answer_gt": "[0.97, 1.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 642 + }, + { + "path": "6432-63722-0021.flac", + "question": "Listen for the precise interval when 'he' is dominant.", + "choice_a": "[2.16, 2.26]", + "choice_b": "[4.39, 4.56]", + "choice_c": "[5.52, 5.96]", + "choice_d": "[1.19, 1.31]", + "answer_gt": "[4.39, 4.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 643 + }, + { + "path": "908-31957-0025.flac", + "question": "Identify the portion of the clip where 'and' is most prominent.", + "choice_a": "[2.78, 2.89]", + "choice_b": "[2.18, 2.75]", + "choice_c": "[7.58, 8.24]", + "choice_d": "[4.06, 14.5]", + "answer_gt": "[4.06, 14.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 644 + }, + { + "path": "6455-66379-0016.flac", + "question": "Identify the portion of the clip where 'changed' is most prominent.", + "choice_a": "[1.29, 1.82]", + "choice_b": "[0.35, 0.76]", + "choice_c": "[0.76, 0.93]", + "choice_d": "[0.93, 1.26]", + "answer_gt": "[0.93, 1.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 645 + }, + { + "path": "6241-61943-0025.flac", + "question": "Spot the exact moment 'the' appears in the audio.", + "choice_a": "[5.11, 5.35]", + "choice_b": "[5.35, 5.48]", + "choice_c": "[0.91, 1.03]", + "choice_d": "[1.32, 1.41]", + "answer_gt": "[0.91, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 646 + }, + { + "path": "4198-12281-0004.flac", + "question": "Pinpoint the passage where 'sanct' is heard in the background.", + "choice_a": "[3.22, 3.51]", + "choice_b": "[1.32, 1.70]", + "choice_c": "[3.10, 3.22]", + "choice_d": "[1.21, 1.32]", + "answer_gt": "[1.32, 1.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 647 + }, + { + "path": "7601-101622-0000_0.flac", + "question": "Listen to this audio clip: when does 'sometimes' occur?", + "choice_a": "[1.59, 22.1]", + "choice_b": "[4.12, 4.46]", + "choice_c": "[6.26, 26.5]", + "choice_d": "[7.92, 18.5]", + "answer_gt": "[7.92, 18.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 648 + }, + { + "path": "6455-67804-0035.flac", + "question": "Locate the part of the audio where 'and' is expressed.", + "choice_a": "[2.27, 2.38]", + "choice_b": "[4.14, 4.45]", + "choice_c": "[4.01, 4.14]", + "choice_d": "[1.83, 2.27]", + "answer_gt": "[2.27, 2.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 649 + }, + { + "path": "8254-84205-0006.flac", + "question": "Identify the moment 'did' is mentioned in the audio.", + "choice_a": "[0.42, 0.55]", + "choice_b": "[1.17, 1.35]", + "choice_c": "[0.55, 0.73]", + "choice_d": "[0.82, 1.17]", + "answer_gt": "[0.55, 0.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 650 + }, + { + "path": "7018-75788-0008.flac", + "question": "When does the audio feature the sound of 'shahrazad'?", + "choice_a": "[0.28, 0.38]", + "choice_b": "[2.45, 3.28]", + "choice_c": "[5.95, 6.47]", + "choice_d": "[3.35, 3.92]", + "answer_gt": "[2.45, 3.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 651 + }, + { + "path": "1988-147956-0011.flac", + "question": "Discover when 'was' is introduced in the conversation.", + "choice_a": "[3.03, 3.19]", + "choice_b": "[0.72, 0.87]", + "choice_c": "[2.04, 2.19]", + "choice_d": "[2.40, 2.86]", + "answer_gt": "[0.72, 0.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 652 + }, + { + "path": "6128-63244-0006.flac", + "question": "At what time interval does 'she' occur in the audio clip?", + "choice_a": "[0.53, 0.74]", + "choice_b": "[1.43, 1.53]", + "choice_c": "[0.74, 0.98]", + "choice_d": "[0.98, 1.06]", + "answer_gt": "[0.53, 0.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 653 + }, + { + "path": "4515-11057-0049.flac", + "question": "Listen to this audio clip; identify the time segment when 'better' is spoken.", + "choice_a": "[2.17, 2.34]", + "choice_b": "[1.13, 1.60]", + "choice_c": "[0.56, 0.70]", + "choice_d": "[0.76, 1.13]", + "answer_gt": "[0.76, 1.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 654 + }, + { + "path": "6267-65525-0004.flac", + "question": "Identify the portion of the clip where 'and' is most prominent.", + "choice_a": "[1.44, 1.59]", + "choice_b": "[1.87, 2.45]", + "choice_c": "[2.65, 2.76]", + "choice_d": "[1.59, 1.87]", + "answer_gt": "[1.44, 1.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 655 + }, + { + "path": "1919-142785-0055.flac", + "question": "What is the timestamp for the initial occurrence of 'follow'?", + "choice_a": "[4.59, 4.74]", + "choice_b": "[7.63, 7.73]", + "choice_c": "[0.77, 1.03]", + "choice_d": "[6.42, 6.83]", + "answer_gt": "[6.42, 6.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 656 + }, + { + "path": "5683-32865-0013.flac", + "question": "Locate the segment where 'you' is faintly heard.", + "choice_a": "[0.19, 0.28]", + "choice_b": "[3.69, 4.03]", + "choice_c": "[0.28, 0.49]", + "choice_d": "[5.95, 6.14]", + "answer_gt": "[0.28, 0.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 657 + }, + { + "path": "8280-266249-0043.flac", + "question": "In the audio track, when is 'go' first noticed?", + "choice_a": "[4.24, 4.42]", + "choice_b": "[2.43, 2.82]", + "choice_c": "[0.82, 1.29]", + "choice_d": "[1.42, 1.74]", + "answer_gt": "[0.82, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 658 + }, + { + "path": "5142-33396-0029.flac", + "question": "Pinpoint the exact moment 'the' appears.", + "choice_a": "[1.11, 1.27]", + "choice_b": "[0.64, 0.71]", + "choice_c": "[1.27, 1.37]", + "choice_d": "[0.71, 1.11]", + "answer_gt": "[0.64, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 659 + }, + { + "path": "4077-13751-0017.flac", + "question": "At what point in the audio is 'record' clearly audible?", + "choice_a": "[0.73, 1.14]", + "choice_b": "[0.33, 0.47]", + "choice_c": "[0.70, 0.73]", + "choice_d": "[0.70, 0.73]", + "answer_gt": "[0.73, 1.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 660 + }, + { + "path": "5683-32866-0008.flac", + "question": "Discover when 'good' is introduced in the conversation.", + "choice_a": "[1.30, 1.79]", + "choice_b": "[0.82, 1.11]", + "choice_c": "[0.74, 0.82]", + "choice_d": "[1.11, 1.30]", + "answer_gt": "[1.11, 1.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 661 + }, + { + "path": "4852-28311-0020.flac", + "question": "Pinpoint the exact moment 'reached' appears.", + "choice_a": "[1.55, 1.87]", + "choice_b": "[1.43, 1.55]", + "choice_c": "[0.45, 0.79]", + "choice_d": "[0.79, 0.93]", + "answer_gt": "[1.55, 1.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 662 + }, + { + "path": "8224-274381-0015.flac", + "question": "Listen to this audio clip; identify the time segment when 'by' is spoken.", + "choice_a": "[0.54, 0.66]", + "choice_b": "[0.76, 11.1]", + "choice_c": "[0.56, 10.7]", + "choice_d": "[3.39, 3.62]", + "answer_gt": "[0.56, 10.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 663 + }, + { + "path": "2412-153954-0014.flac", + "question": "At what time interval does 'no' occur in the audio clip?", + "choice_a": "[2.57, 12.7]", + "choice_b": "[3.24, 3.44]", + "choice_c": "[1.99, 12.1]", + "choice_d": "[9.20, 9.33]", + "answer_gt": "[2.57, 12.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 664 + }, + { + "path": "2830-3980-0007.flac", + "question": "When does the speaker mention 'for'?", + "choice_a": "[0.74, 1.26]", + "choice_b": "[2.34, 2.50]", + "choice_c": "[2.50, 2.61]", + "choice_d": "[8.23, 8.41]", + "answer_gt": "[2.34, 2.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 665 + }, + { + "path": "908-157963-0016.flac", + "question": "Find out the time when 'complain' is immediately occured.", + "choice_a": "[2.01, 2.79]", + "choice_b": "[0.30, 0.46]", + "choice_c": "[3.05, 3.22]", + "choice_d": "[3.22, 3.33]", + "answer_gt": "[2.01, 2.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 666 + }, + { + "path": "4446-2273-0005.flac", + "question": "Determine the time when 'a' happens within the clip.", + "choice_a": "[0.72, 0.75]", + "choice_b": "[1.21, 1.38]", + "choice_c": "[2.00, 2.16]", + "choice_d": "[0.30, 0.36]", + "answer_gt": "[0.72, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 667 + }, + { + "path": "4198-61336-0004.flac", + "question": "Pinpoint the exact moment 'rival' is repeated in the audio.", + "choice_a": "[5.77, 6.16]", + "choice_b": "[2.24, 2.36]", + "choice_c": "[2.51, 3.03]", + "choice_d": "[1.86, 2.08]", + "answer_gt": "[5.77, 6.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 668 + }, + { + "path": "6455-67803-0021.flac", + "question": "When is 'lucia's' first mentioned in the clip?", + "choice_a": "[0.46, 0.54]", + "choice_b": "[0.68, 0.85]", + "choice_c": "[1.77, 2.27]", + "choice_d": "[0.54, 0.68]", + "answer_gt": "[1.77, 2.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 669 + }, + { + "path": "3752-4943-0024.flac", + "question": "When is 'just' first mentioned in the clip?", + "choice_a": "[2.32, 2.66]", + "choice_b": "[1.24, 1.73]", + "choice_c": "[0.72, 0.79]", + "choice_d": "[0.25, 0.59]", + "answer_gt": "[0.25, 0.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 670 + }, + { + "path": "5105-28241-0006.flac", + "question": "Recognize the segment where 'work' is spoken by the speaker.", + "choice_a": "[0.43, 0.50]", + "choice_b": "[0.43, 0.50]", + "choice_c": "[4.12, 4.43]", + "choice_d": "[2.80, 2.91]", + "answer_gt": "[4.12, 4.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 671 + }, + { + "path": "6070-86745-0015.flac", + "question": "When is 'about' first mentioned in the clip?", + "choice_a": "[0.43, 0.93]", + "choice_b": "[0.16, 0.43]", + "choice_c": "[1.34, 1.39]", + "choice_d": "[1.34, 1.39]", + "answer_gt": "[0.16, 0.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 672 + }, + { + "path": "3538-142836-0024.flac", + "question": "Note the time interval that includes the mention of 'desserts'.", + "choice_a": "[2.93, 3.04]", + "choice_b": "[1.13, 1.30]", + "choice_c": "[0.56, 1.09]", + "choice_d": "[0.43, 0.56]", + "answer_gt": "[0.56, 1.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 673 + }, + { + "path": "1462-170142-0021.flac", + "question": "At what interval does 'i' appear in the audio?", + "choice_a": "[6.86, 6.93]", + "choice_b": "[2.15, 2.30]", + "choice_c": "[4.85, 4.92]", + "choice_d": "[4.54, 4.63]", + "answer_gt": "[4.54, 4.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 674 + }, + { + "path": "3081-166546-0055.flac", + "question": "Highlight when 'who' is spoken.", + "choice_a": "[1.69, 1.83]", + "choice_b": "[1.83, 2.02]", + "choice_c": "[1.34, 1.69]", + "choice_d": "[2.39, 2.55]", + "answer_gt": "[1.69, 1.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 675 + }, + { + "path": "5895-34615-0003.flac", + "question": "Highlight when 'industrious' is spoken.", + "choice_a": "[1.14, 1.36]", + "choice_b": "[2.57, 3.31]", + "choice_c": "[4.03, 4.16]", + "choice_d": "[3.31, 4.03]", + "answer_gt": "[2.57, 3.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 676 + }, + { + "path": "8297-275156-0013.flac", + "question": "Discover when 'perhaps' is introduced in the conversation.", + "choice_a": "[4.94, 5.20]", + "choice_b": "[3.61, 3.66]", + "choice_c": "[2.97, 3.39]", + "choice_d": "[1.15, 1.82]", + "answer_gt": "[1.15, 1.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 677 + }, + { + "path": "2033-164916-0004.flac", + "question": "Listen for when 'do' is emphasized by the narrator.", + "choice_a": "[2.65, 2.84]", + "choice_b": "[3.99, 4.18]", + "choice_c": "[4.50, 4.76]", + "choice_d": "[4.76, 4.98]", + "answer_gt": "[4.50, 4.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 678 + }, + { + "path": "5442-41169-0006.flac", + "question": "At what interval does 'at' appear in the audio?", + "choice_a": "[1.01, 1.15]", + "choice_b": "[0.42, 0.77]", + "choice_c": "[1.69, 2.19]", + "choice_d": "[1.24, 1.69]", + "answer_gt": "[1.01, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 679 + }, + { + "path": "3081-166546-0015.flac", + "question": "When is 'i' first heard in the recording?", + "choice_a": "[1.55, 1.77]", + "choice_b": "[0.68, 0.88]", + "choice_c": "[0.88, 1.16]", + "choice_d": "[0.57, 0.68]", + "answer_gt": "[0.57, 0.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 680 + }, + { + "path": "4077-13751-0020.flac", + "question": "Note the time when 'brutality' is first mentioned in the audio.", + "choice_a": "[2.24, 2.78]", + "choice_b": "[0.73, 0.80]", + "choice_c": "[2.86, 3.54]", + "choice_d": "[3.74, 4.37]", + "answer_gt": "[2.86, 3.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 681 + }, + { + "path": "2506-11278-0026.flac", + "question": "Recognize the time when 'life' is mentioned by the speaker.", + "choice_a": "[2.88, 3.24]", + "choice_b": "[5.16, 5.42]", + "choice_c": "[0.36, 0.84]", + "choice_d": "[5.54, 5.66]", + "answer_gt": "[2.88, 3.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 682 + }, + { + "path": "8463-294828-0021.flac", + "question": "Determine the time when 'route' happens within the clip.", + "choice_a": "[1.20, 1.50]", + "choice_b": "[0.22, 0.32]", + "choice_c": "[0.32, 0.72]", + "choice_d": "[1.91, 2.26]", + "answer_gt": "[0.32, 0.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 683 + }, + { + "path": "8188-269288-0032.flac", + "question": "Ascertain the exact time 'by' is mentioned in the audio.", + "choice_a": "[1.20, 1.42]", + "choice_b": "[0.78, 0.98]", + "choice_c": "[1.62, 1.76]", + "choice_d": "[0.98, 1.12]", + "answer_gt": "[1.62, 1.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 684 + }, + { + "path": "3853-163249-0032.flac", + "question": "Pinpoint the passage where 'you' is heard in the background.", + "choice_a": "[2.17, 2.34]", + "choice_b": "[3.68, 4.21]", + "choice_c": "[0.32, 0.43]", + "choice_d": "[1.09, 1.44]", + "answer_gt": "[0.32, 0.43]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 685 + }, + { + "path": "7975-280084-0003.flac", + "question": "Listen for the occurrence of 'allen' in the recording.", + "choice_a": "[2.47, 2.70]", + "choice_b": "[0.69, 1.07]", + "choice_c": "[4.59, 4.73]", + "choice_d": "[2.76, 2.91]", + "answer_gt": "[0.69, 1.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 686 + }, + { + "path": "1688-142285-0073.flac", + "question": "When does 'foreshadowed' occur in the audio?", + "choice_a": "[1.87, 2.51]", + "choice_b": "[5.54, 5.99]", + "choice_c": "[2.89, 3.04]", + "choice_d": "[2.60, 2.89]", + "answer_gt": "[1.87, 2.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 687 + }, + { + "path": "1995-1836-0009.flac", + "question": "Find the moment 'capacity' is clearly audible in the audio.", + "choice_a": "[0.13, 0.46]", + "choice_b": "[1.22, 1.59]", + "choice_c": "[2.81, 3.54]", + "choice_d": "[0.62, 1.18]", + "answer_gt": "[2.81, 3.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 688 + }, + { + "path": "7127-75947-0015.flac", + "question": "Discover when 'returned' is introduced in the conversation.", + "choice_a": "[2.15, 2.27]", + "choice_b": "[1.61, 1.72]", + "choice_c": "[5.08, 5.21]", + "choice_d": "[4.24, 4.78]", + "answer_gt": "[4.24, 4.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 689 + }, + { + "path": "1320-122617-0015.flac", + "question": "Determine the time when 'instead' happens within the clip.", + "choice_a": "[0.36, 0.42]", + "choice_b": "[1.35, 1.94]", + "choice_c": "[3.21, 3.35]", + "choice_d": "[0.84, 1.24]", + "answer_gt": "[0.84, 1.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 690 + }, + { + "path": "2428-83699-0006.flac", + "question": "Identify when 'do' is heard in the audio.", + "choice_a": "[1.08, 1.65]", + "choice_b": "[5.83, 5.95]", + "choice_c": "[3.03, 3.80]", + "choice_d": "[6.16, 6.27]", + "answer_gt": "[5.83, 5.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 691 + }, + { + "path": "1651-136854-0013.flac", + "question": "Identify the moment 'scholar's' is mentioned in the audio.", + "choice_a": "[0.41, 0.94]", + "choice_b": "[0.94, 1.43]", + "choice_c": "[0.94, 1.43]", + "choice_d": "[0.94, 1.43]", + "answer_gt": "[0.41, 0.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 692 + }, + { + "path": "6267-65525-0013.flac", + "question": "Note the time when 'that' is first mentioned in the audio.", + "choice_a": "[0.77, 1.15]", + "choice_b": "[3.75, 3.94]", + "choice_c": "[7.43, 7.83]", + "choice_d": "[2.14, 2.51]", + "answer_gt": "[3.75, 3.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 693 + }, + { + "path": "6241-61946-0022.flac", + "question": "Recognize the segment where 'when' is spoken by the speaker.", + "choice_a": "[3.90, 4.33]", + "choice_b": "[2.19, 2.62]", + "choice_c": "[6.95, 6.98]", + "choice_d": "[5.13, 5.27]", + "answer_gt": "[5.13, 5.27]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 694 + }, + { + "path": "8173-294714-0044.flac", + "question": "Listen to this audio clip: when does 'of' occur?", + "choice_a": "[2.45, 2.51]", + "choice_b": "[0.79, 0.91]", + "choice_c": "[0.35, 0.48]", + "choice_d": "[0.52, 0.79]", + "answer_gt": "[0.79, 0.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 695 + }, + { + "path": "2078-142845-0026.flac", + "question": "Pinpoint the exact moment 'make' appears.", + "choice_a": "[0.53, 0.64]", + "choice_b": "[1.12, 1.43]", + "choice_c": "[1.12, 1.43]", + "choice_d": "[0.64, 1.07]", + "answer_gt": "[0.64, 1.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 696 + }, + { + "path": "5895-34629-0025.flac", + "question": "Determine the time when 'was' happens within the clip.", + "choice_a": "[1.38, 1.78]", + "choice_b": "[3.75, 3.85]", + "choice_c": "[0.61, 1.20]", + "choice_d": "[1.20, 1.38]", + "answer_gt": "[1.20, 1.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 697 + }, + { + "path": "5683-32866-0029.flac", + "question": "Pinpoint the exact moment 'somehow' appears.", + "choice_a": "[0.28, 0.84]", + "choice_b": "[0.84, 0.94]", + "choice_c": "[1.09, 1.40]", + "choice_d": "[1.09, 1.40]", + "answer_gt": "[0.28, 0.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 698 + }, + { + "path": "84-121550-0022.flac", + "question": "Pinpoint the passage where 'i' is heard in the background.", + "choice_a": "[3.44, 3.66]", + "choice_b": "[6.89, 7.08]", + "choice_c": "[1.25, 1.41]", + "choice_d": "[6.31, 6.42]", + "answer_gt": "[6.31, 6.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 699 + }, + { + "path": "1462-170142-0028.flac", + "question": "When is 'it' first heard in the recording?", + "choice_a": "[0.96, 1.08]", + "choice_b": "[1.96, 2.24]", + "choice_c": "[1.08, 1.37]", + "choice_d": "[1.37, 1.67]", + "answer_gt": "[0.96, 1.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 700 + }, + { + "path": "4323-18416-0012.flac", + "question": "Listen for the occurrence of 'is' in the recording.", + "choice_a": "[0.86, 1.05]", + "choice_b": "[1.23, 1.37]", + "choice_c": "[1.05, 1.23]", + "choice_d": "[0.74, 0.86]", + "answer_gt": "[1.23, 1.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 701 + }, + { + "path": "7105-2340-0011.flac", + "question": "Determine the point in the audio when 'the' is repeated.", + "choice_a": "[3.04, 3.20]", + "choice_b": "[2.32, 2.77]", + "choice_c": "[2.06, 2.32]", + "choice_d": "[0.39, 0.53]", + "answer_gt": "[0.39, 0.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 702 + }, + { + "path": "7850-286674-0014.flac", + "question": "Find out the time when 'the' is immediately occured.", + "choice_a": "[4.96, 5.40]", + "choice_b": "[4.58, 4.93]", + "choice_c": "[4.33, 4.58]", + "choice_d": "[2.47, 2.58]", + "answer_gt": "[2.47, 2.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 703 + }, + { + "path": "8555-284449-0010.flac", + "question": "At what point in the audio is 'will' clearly audible?", + "choice_a": "[0.24, 0.37]", + "choice_b": "[0.65, 1.12]", + "choice_c": "[1.12, 1.19]", + "choice_d": "[0.37, 0.50]", + "answer_gt": "[0.37, 0.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 704 + }, + { + "path": "2902-9006-0017.flac", + "question": "Pinpoint the exact moment 'last' appears.", + "choice_a": "[4.54, 14.9]", + "choice_b": "[5.35, 15.5]", + "choice_c": "[2.12, 2.19]", + "choice_d": "[3.86, 4.10]", + "answer_gt": "[4.54, 14.9]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 705 + }, + { + "path": "1688-142285-0005.flac", + "question": "Identify when 'always' is heard in the audio.", + "choice_a": "[2.00, 2.30]", + "choice_b": "[2.44, 2.60]", + "choice_c": "[1.12, 1.49]", + "choice_d": "[2.60, 3.06]", + "answer_gt": "[1.12, 1.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 706 + }, + { + "path": "5895-34622-0023.flac", + "question": "Identify the portion of the clip where 'in' is most prominent.", + "choice_a": "[0.94, 1.04]", + "choice_b": "[0.94, 1.04]", + "choice_c": "[2.25, 2.32]", + "choice_d": "[0.94, 1.04]", + "answer_gt": "[0.94, 1.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 707 + }, + { + "path": "6295-244435-0006.flac", + "question": "Find out the time when 'soon' is immediately occured.", + "choice_a": "[0.31, 0.37]", + "choice_b": "[1.32, 1.44]", + "choice_c": "[0.37, 1.32]", + "choice_d": "[1.44, 1.79]", + "answer_gt": "[1.44, 1.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 708 + }, + { + "path": "2035-147961-0033.flac", + "question": "What is the timestamp for when 'of' occurs?", + "choice_a": "[0.43, 0.61]", + "choice_b": "[0.27, 0.43]", + "choice_c": "[0.99, 1.10]", + "choice_d": "[0.78, 0.99]", + "answer_gt": "[0.99, 1.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 709 + }, + { + "path": "4570-14911-0002.flac", + "question": "Discover when 'down' is introduced in the conversation.", + "choice_a": "[1.79, 2.07]", + "choice_b": "[2.31, 2.76]", + "choice_c": "[1.47, 1.58]", + "choice_d": "[6.42, 6.52]", + "answer_gt": "[1.79, 2.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 710 + }, + { + "path": "61-70968-0037.flac", + "question": "Locate the segment where 'name' is faintly heard.", + "choice_a": "[2.53, 2.78]", + "choice_b": "[1.11, 1.49]", + "choice_c": "[0.98, 1.11]", + "choice_d": "[2.46, 2.53]", + "answer_gt": "[1.11, 1.49]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 711 + }, + { + "path": "7018-75789-0019.flac", + "question": "Determine the point in the audio when 'dominion' is discussed.", + "choice_a": "[5.54, 5.81]", + "choice_b": "[0.27, 0.51]", + "choice_c": "[2.96, 3.52]", + "choice_d": "[3.75, 4.32]", + "answer_gt": "[3.75, 4.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 712 + }, + { + "path": "3536-23268-0019.flac", + "question": "When does the audio feature the sound of 'of'?", + "choice_a": "[3.35, 3.59]", + "choice_b": "[8.67, 9.19]", + "choice_c": "[5.38, 5.48]", + "choice_d": "[3.77, 4.01]", + "answer_gt": "[5.38, 5.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 713 + }, + { + "path": "2277-149897-0017.flac", + "question": "Pinpoint the exact moment 'or' is repeated in the audio.", + "choice_a": "[2.48, 2.57]", + "choice_b": "[2.97, 3.13]", + "choice_c": "[0.55, 0.66]", + "choice_d": "[3.41, 3.53]", + "answer_gt": "[3.41, 3.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 714 + }, + { + "path": "4446-2273-0006.flac", + "question": "In the audio track, when is 'sketches' first noticed?", + "choice_a": "[1.62, 1.86]", + "choice_b": "[1.31, 1.57]", + "choice_c": "[0.29, 0.42]", + "choice_d": "[0.57, 1.10]", + "answer_gt": "[0.57, 1.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 715 + }, + { + "path": "4852-28312-0023.flac", + "question": "When is 'warehouses' first mentioned in the clip?", + "choice_a": "[1.15, 1.49]", + "choice_b": "[0.36, 1.03]", + "choice_c": "[1.03, 1.15]", + "choice_d": "[1.03, 1.15]", + "answer_gt": "[0.36, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 716 + }, + { + "path": "3081-166546-0010.flac", + "question": "Determine the time when 'i' happens within the clip.", + "choice_a": "[0.12, 0.15]", + "choice_b": "[3.65, 4.03]", + "choice_c": "[0.49, 0.91]", + "choice_d": "[1.45, 1.99]", + "answer_gt": "[0.12, 0.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 717 + }, + { + "path": "260-123286-0031.flac", + "question": "Listen to this audio clip: when does 'ichthyosaurus' occur?", + "choice_a": "[0.57, 0.81]", + "choice_b": "[1.08, 2.12]", + "choice_c": "[3.75, 4.29]", + "choice_d": "[3.56, 3.66]", + "answer_gt": "[1.08, 2.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 718 + }, + { + "path": "7021-79740-0007.flac", + "question": "Locate the segment where 'to' is faintly heard.", + "choice_a": "[3.45, 3.51]", + "choice_b": "[3.51, 3.90]", + "choice_c": "[3.45, 3.51]", + "choice_d": "[1.46, 1.58]", + "answer_gt": "[1.46, 1.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 719 + }, + { + "path": "5484-24318-0013.flac", + "question": "Listen to this audio clip: when does 'general' occur?", + "choice_a": "[7.28, 7.76]", + "choice_b": "[4.22, 4.28]", + "choice_c": "[0.76, 10.9]", + "choice_d": "[1.43, 1.85]", + "answer_gt": "[7.28, 7.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 720 + }, + { + "path": "4831-29134-0002.flac", + "question": "Find out the time when 'to' is immediately occured.", + "choice_a": "[5.01, 5.17]", + "choice_b": "[0.62, 0.90]", + "choice_c": "[1.35, 1.46]", + "choice_d": "[4.35, 4.71]", + "answer_gt": "[1.35, 1.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 721 + }, + { + "path": "6345-64257-0018.flac", + "question": "Ascertain the point at which 'into' enters the discussion.", + "choice_a": "[1.77, 1.92]", + "choice_b": "[0.84, 1.00]", + "choice_c": "[0.56, 0.84]", + "choice_d": "[1.58, 1.77]", + "answer_gt": "[1.58, 1.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 722 + }, + { + "path": "3764-168670-0011.flac", + "question": "What is the duration of 'him' being mentioned in the conversation?", + "choice_a": "[1.56, 1.67]", + "choice_b": "[3.71, 3.78]", + "choice_c": "[2.17, 2.24]", + "choice_d": "[1.13, 1.56]", + "answer_gt": "[1.56, 1.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 723 + }, + { + "path": "422-122949-0032.flac", + "question": "When is 'probably' first heard in the recording?", + "choice_a": "[0.53, 1.06]", + "choice_b": "[2.14, 12.3]", + "choice_c": "[2.33, 12.5]", + "choice_d": "[9.71, 9.88]", + "answer_gt": "[0.53, 1.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 724 + }, + { + "path": "8173-294714-0009.flac", + "question": "At what interval does 'you' appear in the audio?", + "choice_a": "[1.17, 1.60]", + "choice_b": "[1.02, 1.17]", + "choice_c": "[0.94, 1.02]", + "choice_d": "[0.59, 0.94]", + "answer_gt": "[1.02, 1.17]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 725 + }, + { + "path": "3764-168670-0003.flac", + "question": "Determine the time when 'lady's' happens within the clip.", + "choice_a": "[3.01, 3.19]", + "choice_b": "[1.40, 1.44]", + "choice_c": "[1.44, 1.87]", + "choice_d": "[3.39, 3.51]", + "answer_gt": "[1.44, 1.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 726 + }, + { + "path": "5895-34622-0013.flac", + "question": "At what point in the audio is 'dreams' clearly audible?", + "choice_a": "[8.18, 8.35]", + "choice_b": "[6.32, 6.67]", + "choice_c": "[6.67, 7.40]", + "choice_d": "[9.08, 9.61]", + "answer_gt": "[6.32, 6.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 727 + }, + { + "path": "3764-168670-0054.flac", + "question": "Discover when 'around' is introduced in the conversation.", + "choice_a": "[4.71, 4.91]", + "choice_b": "[0.30, 0.41]", + "choice_c": "[3.13, 3.23]", + "choice_d": "[2.82, 3.13]", + "answer_gt": "[2.82, 3.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 728 + }, + { + "path": "1650-167613-0024.flac", + "question": "Determine the precise moment 'why' in the audio's pace.", + "choice_a": "[1.79, 1.90]", + "choice_b": "[0.81, 1.14]", + "choice_c": "[1.21, 1.72]", + "choice_d": "[0.26, 0.54]", + "answer_gt": "[0.26, 0.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 729 + }, + { + "path": "4323-55228-0050.flac", + "question": "Recognize the time when 'nodded' is most clearly articulated.", + "choice_a": "[1.22, 1.45]", + "choice_b": "[0.91, 1.22]", + "choice_c": "[0.39, 0.85]", + "choice_d": "[0.39, 0.85]", + "answer_gt": "[0.91, 1.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 730 + }, + { + "path": "84-121123-0003.flac", + "question": "Note the time interval that includes the mention of 'its'.", + "choice_a": "[5.60, 5.81]", + "choice_b": "[4.17, 4.26]", + "choice_c": "[1.73, 1.89]", + "choice_d": "[3.11, 3.26]", + "answer_gt": "[5.60, 5.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 731 + }, + { + "path": "7902-96595-0006.flac", + "question": "Locate the segment where 'stand' is faintly heard.", + "choice_a": "[2.81, 3.10]", + "choice_b": "[0.51, 0.73]", + "choice_c": "[0.41, 0.51]", + "choice_d": "[0.41, 0.51]", + "answer_gt": "[2.81, 3.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 732 + }, + { + "path": "3081-166546-0002.flac", + "question": "Identify when 'a' is heard in the audio.", + "choice_a": "[0.42, 0.96]", + "choice_b": "[1.54, 2.32]", + "choice_c": "[1.17, 1.22]", + "choice_d": "[1.22, 1.54]", + "answer_gt": "[1.17, 1.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 733 + }, + { + "path": "4572-112381-0017.flac", + "question": "When does the audio feature the sound of 'fondly'?", + "choice_a": "[4.31, 4.47]", + "choice_b": "[2.06, 2.70]", + "choice_c": "[2.89, 2.99]", + "choice_d": "[0.66, 1.15]", + "answer_gt": "[0.66, 1.15]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 734 + }, + { + "path": "6128-63244-0002.flac", + "question": "Find out the time when 'them' is immediately occured.", + "choice_a": "[0.40, 0.76]", + "choice_b": "[1.38, 1.81]", + "choice_c": "[0.76, 0.87]", + "choice_d": "[1.27, 1.38]", + "answer_gt": "[1.38, 1.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 735 + }, + { + "path": "260-123288-0019.flac", + "question": "Find the duration when 'at' occurs consecutively in the audio.", + "choice_a": "[1.66, 2.10]", + "choice_b": "[0.55, 0.72]", + "choice_c": "[1.05, 1.11]", + "choice_d": "[1.11, 1.51]", + "answer_gt": "[0.55, 0.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 736 + }, + { + "path": "6313-76958-0005.flac", + "question": "Find the moment 'around' is clearly audible in the audio.", + "choice_a": "[3.35, 3.59]", + "choice_b": "[2.59, 2.88]", + "choice_c": "[3.65, 3.88]", + "choice_d": "[4.20, 4.36]", + "answer_gt": "[3.35, 3.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 737 + }, + { + "path": "777-126732-0038.flac", + "question": "Note the time when 'he' is first mentioned in the audio.", + "choice_a": "[1.66, 2.07]", + "choice_b": "[0.57, 1.12]", + "choice_c": "[1.12, 1.24]", + "choice_d": "[0.44, 0.57]", + "answer_gt": "[0.44, 0.57]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 738 + }, + { + "path": "121-127105-0020.flac", + "question": "Ascertain the exact time 'any' is mentioned in the audio.", + "choice_a": "[6.57, 7.00]", + "choice_b": "[0.85, 1.12]", + "choice_c": "[1.39, 11.6]", + "choice_d": "[9.93, 10.2]", + "answer_gt": "[1.39, 11.6]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 739 + }, + { + "path": "1988-147956-0021.flac", + "question": "Recognize the time when 'her' is most clearly articulated.", + "choice_a": "[0.64, 0.88]", + "choice_b": "[1.56, 1.64]", + "choice_c": "[3.12, 3.44]", + "choice_d": "[2.83, 3.12]", + "answer_gt": "[1.56, 1.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 740 + }, + { + "path": "4515-11057-0110.flac", + "question": "When is 'received' first heard in the recording?", + "choice_a": "[0.48, 0.95]", + "choice_b": "[1.78, 2.38]", + "choice_c": "[0.95, 1.38]", + "choice_d": "[1.60, 1.78]", + "answer_gt": "[1.78, 2.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 741 + }, + { + "path": "1188-133604-0002.flac", + "question": "When does 'of' occur in the audio?", + "choice_a": "[3.81, 3.93]", + "choice_b": "[4.81, 15.2]", + "choice_c": "[1.39, 1.50]", + "choice_d": "[4.06, 4.14]", + "answer_gt": "[1.39, 1.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 742 + }, + { + "path": "1701-141760-0006.flac", + "question": "Identify when 'out' is heard in the audio.", + "choice_a": "[2.00, 2.14]", + "choice_b": "[1.32, 1.49]", + "choice_c": "[1.65, 1.77]", + "choice_d": "[2.14, 2.25]", + "answer_gt": "[2.00, 2.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 743 + }, + { + "path": "2035-152373-0012.flac", + "question": "When does 'to' occur in the audio?", + "choice_a": "[3.91, 4.03]", + "choice_b": "[3.91, 4.03]", + "choice_c": "[7.85, 8.55]", + "choice_d": "[6.31, 7.03]", + "answer_gt": "[3.91, 4.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 744 + }, + { + "path": "4515-11057-0050.flac", + "question": "Recognize the segment where 'brute' is spoken by the speaker.", + "choice_a": "[1.46, 1.56]", + "choice_b": "[0.86, 0.98]", + "choice_c": "[1.04, 1.46]", + "choice_d": "[0.98, 1.04]", + "answer_gt": "[1.04, 1.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 745 + }, + { + "path": "251-137823-0000.flac", + "question": "When is 'to' first heard in the recording?", + "choice_a": "[1.19, 1.53]", + "choice_b": "[1.12, 1.19]", + "choice_c": "[0.66, 0.78]", + "choice_d": "[2.28, 2.38]", + "answer_gt": "[1.12, 1.19]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 746 + }, + { + "path": "1630-141772-0019.flac", + "question": "In the audio track, when is 'he' first noticed?", + "choice_a": "[0.76, 11.0]", + "choice_b": "[3.82, 14.2]", + "choice_c": "[1.05, 11.5]", + "choice_d": "[5.85, 16.3]", + "answer_gt": "[1.05, 11.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 747 + }, + { + "path": "6841-88291-0011.flac", + "question": "At what interval does 'among' appear in the audio?", + "choice_a": "[2.44, 2.62]", + "choice_b": "[2.72, 3.02]", + "choice_c": "[1.09, 1.22]", + "choice_d": "[0.29, 0.63]", + "answer_gt": "[2.72, 3.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 748 + }, + { + "path": "3528-168669-0016.flac", + "question": "When does the audio feature the sound of 'is'?", + "choice_a": "[0.53, 0.65]", + "choice_b": "[1.43, 1.69]", + "choice_c": "[0.65, 0.83]", + "choice_d": "[2.01, 2.14]", + "answer_gt": "[0.65, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 749 + }, + { + "path": "121-127105-0032.flac", + "question": "Note the time interval that includes the mention of 'yes'.", + "choice_a": "[0.34, 0.82]", + "choice_b": "[1.04, 1.31]", + "choice_c": "[1.65, 2.14]", + "choice_d": "[2.14, 2.27]", + "answer_gt": "[0.34, 0.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 750 + }, + { + "path": "8188-269288-0039.flac", + "question": "Pinpoint the exact moment 'i' is repeated in the audio.", + "choice_a": "[0.62, 0.91]", + "choice_b": "[0.52, 0.62]", + "choice_c": "[1.22, 1.54]", + "choice_d": "[1.79, 2.10]", + "answer_gt": "[0.52, 0.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 751 + }, + { + "path": "1585-157660-0012.flac", + "question": "Determine the time when 'i' happens within the clip.", + "choice_a": "[0.48, 0.62]", + "choice_b": "[3.34, 3.74]", + "choice_c": "[3.74, 4.06]", + "choice_d": "[3.19, 3.34]", + "answer_gt": "[0.48, 0.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 752 + }, + { + "path": "116-288048-0010.flac", + "question": "When is 'companions' first heard in the recording?", + "choice_a": "[2.47, 2.64]", + "choice_b": "[0.91, 1.48]", + "choice_c": "[4.46, 4.60]", + "choice_d": "[0.42, 0.52]", + "answer_gt": "[0.91, 1.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 753 + }, + { + "path": "2412-153948-0008.flac", + "question": "Spot the exact moment 'accustomed' appears in the audio.", + "choice_a": "[0.47, 0.60]", + "choice_b": "[4.52, 4.64]", + "choice_c": "[3.18, 3.28]", + "choice_d": "[2.34, 3.09]", + "answer_gt": "[2.34, 3.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 754 + }, + { + "path": "7697-245712-0003.flac", + "question": "Find the duration when 'things' occurs consecutively in the audio.", + "choice_a": "[1.33, 1.43]", + "choice_b": "[3.37, 3.56]", + "choice_c": "[1.77, 1.94]", + "choice_d": "[4.65, 4.94]", + "answer_gt": "[4.65, 4.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 755 + }, + { + "path": "3331-159609-0011.flac", + "question": "Establish when 'which' is first mentioned in the clip.", + "choice_a": "[8.26, 8.34]", + "choice_b": "[1.78, 1.84]", + "choice_c": "[5.31, 5.54]", + "choice_d": "[1.78, 1.84]", + "answer_gt": "[5.31, 5.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 756 + }, + { + "path": "8461-281231-0017.flac", + "question": "Determine the precise moment 'of' in the audio's pace.", + "choice_a": "[0.72, 0.83]", + "choice_b": "[3.72, 3.95]", + "choice_c": "[5.45, 5.75]", + "choice_d": "[0.38, 0.72]", + "answer_gt": "[0.72, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 757 + }, + { + "path": "1221-135767-0018.flac", + "question": "When does 'of' occur in the audio?", + "choice_a": "[2.92, 3.05]", + "choice_b": "[8.33, 8.60]", + "choice_c": "[6.22, 6.34]", + "choice_d": "[1.42, 1.51]", + "answer_gt": "[6.22, 6.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 758 + }, + { + "path": "4446-2275-0041.flac", + "question": "Discover when 'i' is introduced in the conversation.", + "choice_a": "[2.07, 2.20]", + "choice_b": "[2.31, 2.52]", + "choice_c": "[2.20, 2.31]", + "choice_d": "[3.25, 3.46]", + "answer_gt": "[2.20, 2.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 759 + }, + { + "path": "5484-24318-0033.flac", + "question": "When does the audio feature the sound of 'before'?", + "choice_a": "[2.79, 3.00]", + "choice_b": "[4.01, 4.36]", + "choice_c": "[2.63, 2.76]", + "choice_d": "[1.22, 1.51]", + "answer_gt": "[1.22, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 760 + }, + { + "path": "1701-141759-0018.flac", + "question": "Ascertain the point at which 'him' enters the discussion.", + "choice_a": "[0.76, 0.93]", + "choice_b": "[1.34, 1.53]", + "choice_c": "[0.93, 1.19]", + "choice_d": "[1.19, 1.34]", + "answer_gt": "[1.34, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 761 + }, + { + "path": "7697-105815-0033.flac", + "question": "At what point in the audio is 'his' clearly audible?", + "choice_a": "[1.67, 1.81]", + "choice_b": "[0.51, 0.77]", + "choice_c": "[2.52, 2.68]", + "choice_d": "[1.81, 1.93]", + "answer_gt": "[1.81, 1.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 762 + }, + { + "path": "6467-94831-0011.flac", + "question": "In the audio track, when is 'face' first noticed?", + "choice_a": "[0.28, 0.55]", + "choice_b": "[2.46, 2.81]", + "choice_c": "[2.07, 2.23]", + "choice_d": "[0.55, 0.87]", + "answer_gt": "[2.46, 2.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 763 + }, + { + "path": "2277-149896-0013.flac", + "question": "Identify the time segment when 'he' is mentioned in the audio.", + "choice_a": "[2.63, 2.94]", + "choice_b": "[2.49, 2.63]", + "choice_c": "[0.06, 0.29]", + "choice_d": "[2.31, 2.49]", + "answer_gt": "[0.06, 0.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 764 + }, + { + "path": "1255-90407-0007.flac", + "question": "In the audio track, when is 'been' first noticed?", + "choice_a": "[1.51, 2.11]", + "choice_b": "[1.41, 1.51]", + "choice_c": "[0.82, 0.98]", + "choice_d": "[3.04, 3.39]", + "answer_gt": "[0.82, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 765 + }, + { + "path": "8555-284449-0013.flac", + "question": "Note the time interval that includes the mention of 'you're'.", + "choice_a": "[2.42, 2.52]", + "choice_b": "[1.32, 1.67]", + "choice_c": "[3.85, 3.98]", + "choice_d": "[0.80, 1.19]", + "answer_gt": "[3.85, 3.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 766 + }, + { + "path": "8288-274162-0036.flac", + "question": "What is the timestamp for when 'what' occurs?", + "choice_a": "[0.52, 0.67]", + "choice_b": "[3.01, 3.36]", + "choice_c": "[4.00, 4.13]", + "choice_d": "[3.69, 3.84]", + "answer_gt": "[3.69, 3.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 767 + }, + { + "path": "61-70968-0006.flac", + "question": "At what time interval does 'as' occur in the audio clip?", + "choice_a": "[2.14, 2.24]", + "choice_b": "[1.22, 1.51]", + "choice_c": "[1.51, 1.63]", + "choice_d": "[0.72, 1.06]", + "answer_gt": "[1.51, 1.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 768 + }, + { + "path": "3764-168670-0029.flac", + "question": "Highlight when 'the' is spoken.", + "choice_a": "[2.71, 2.87]", + "choice_b": "[0.54, 0.61]", + "choice_c": "[1.01, 1.13]", + "choice_d": "[0.54, 0.61]", + "answer_gt": "[0.54, 0.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 769 + }, + { + "path": "700-122868-0029.flac", + "question": "Listen for when 'anne' is emphasized by the narrator.", + "choice_a": "[0.53, 0.66]", + "choice_b": "[1.08, 1.26]", + "choice_c": "[0.66, 0.69]", + "choice_d": "[0.66, 0.69]", + "answer_gt": "[1.08, 1.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 770 + }, + { + "path": "3536-8226-0006.flac", + "question": "When in the audio does 'trevelyan' coincide with a specific event?", + "choice_a": "[5.60, 6.31]", + "choice_b": "[4.11, 4.59]", + "choice_c": "[2.67, 2.75]", + "choice_d": "[2.67, 2.75]", + "answer_gt": "[5.60, 6.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 771 + }, + { + "path": "4831-25894-0011.flac", + "question": "Determine the point in the audio when 'rather' is repeated.", + "choice_a": "[5.22, 5.35]", + "choice_b": "[0.59, 0.88]", + "choice_c": "[0.94, 1.35]", + "choice_d": "[1.35, 1.54]", + "answer_gt": "[0.59, 0.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 772 + }, + { + "path": "3570-5695-0015.flac", + "question": "Determine the precise moment 'perhaps' in the audio's pace.", + "choice_a": "[6.59, 6.87]", + "choice_b": "[5.87, 6.00]", + "choice_c": "[0.28, 0.34]", + "choice_d": "[3.35, 3.71]", + "answer_gt": "[3.35, 3.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 773 + }, + { + "path": "1688-142285-0054.flac", + "question": "Ascertain the exact time 'nicholas' is mentioned in the audio.", + "choice_a": "[1.51, 1.83]", + "choice_b": "[1.83, 2.01]", + "choice_c": "[0.47, 0.87]", + "choice_d": "[1.04, 1.30]", + "answer_gt": "[0.47, 0.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 774 + }, + { + "path": "6241-61946-0018.flac", + "question": "Find out the time when 'own' is immediately occured.", + "choice_a": "[2.04, 2.21]", + "choice_b": "[1.85, 2.00]", + "choice_c": "[2.82, 3.05]", + "choice_d": "[1.74, 1.85]", + "answer_gt": "[2.04, 2.21]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 775 + }, + { + "path": "4294-14317-0007.flac", + "question": "When does the speaker mention 'to'?", + "choice_a": "[5.56, 5.91]", + "choice_b": "[3.46, 4.21]", + "choice_c": "[1.27, 1.39]", + "choice_d": "[2.27, 2.38]", + "answer_gt": "[1.27, 1.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 776 + }, + { + "path": "8131-117016-0046.flac", + "question": "Locate the part of the audio where 'got' is expressed.", + "choice_a": "[2.97, 3.16]", + "choice_b": "[1.17, 1.80]", + "choice_c": "[3.90, 4.06]", + "choice_d": "[0.54, 0.79]", + "answer_gt": "[2.97, 3.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 777 + }, + { + "path": "700-122868-0010.flac", + "question": "When does the audio feature the sound of 'was'?", + "choice_a": "[0.25, 0.37]", + "choice_b": "[1.75, 2.01]", + "choice_c": "[4.00, 4.14]", + "choice_d": "[2.85, 3.16]", + "answer_gt": "[4.00, 4.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 778 + }, + { + "path": "1320-122612-0008.flac", + "question": "When does 'of' occur in the audio?", + "choice_a": "[0.80, 0.90]", + "choice_b": "[0.80, 0.90]", + "choice_c": "[1.64, 2.14]", + "choice_d": "[5.54, 6.17]", + "answer_gt": "[0.80, 0.90]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 779 + }, + { + "path": "8555-284447-0022.flac", + "question": "In the audio track, when is 'a' first noticed?", + "choice_a": "[2.09, 2.37]", + "choice_b": "[0.55, 0.61]", + "choice_c": "[1.58, 1.92]", + "choice_d": "[2.37, 2.47]", + "answer_gt": "[0.55, 0.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 780 + }, + { + "path": "3997-180297-0017.flac", + "question": "When is 'in' first heard in the recording?", + "choice_a": "[0.78, 1.19]", + "choice_b": "[1.38, 1.53]", + "choice_c": "[1.38, 1.53]", + "choice_d": "[1.19, 1.38]", + "answer_gt": "[1.19, 1.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 781 + }, + { + "path": "4507-16021-0056.flac", + "question": "Recognize the time when 'necessarily' is most clearly articulated.", + "choice_a": "[2.25, 2.62]", + "choice_b": "[3.51, 3.77]", + "choice_c": "[3.77, 4.89]", + "choice_d": "[3.29, 3.51]", + "answer_gt": "[3.77, 4.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 782 + }, + { + "path": "7641-96684-0037.flac", + "question": "Find out when the speaker refers to 'have' during the audio.", + "choice_a": "[1.44, 1.70]", + "choice_b": "[1.70, 1.95]", + "choice_c": "[1.95, 2.48]", + "choice_d": "[2.84, 2.95]", + "answer_gt": "[2.84, 2.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 783 + }, + { + "path": "652-130726-0035.flac", + "question": "Highlight when 'oregon' is spoken.", + "choice_a": "[2.48, 3.27]", + "choice_b": "[3.95, 4.38]", + "choice_c": "[1.79, 1.97]", + "choice_d": "[1.97, 2.16]", + "answer_gt": "[3.95, 4.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 784 + }, + { + "path": "777-126732-0066.flac", + "question": "Determine the time when 'did' happens within the clip.", + "choice_a": "[9.68, 9.82]", + "choice_b": "[3.62, 3.79]", + "choice_c": "[8.09, 8.34]", + "choice_d": "[3.79, 4.27]", + "answer_gt": "[9.68, 9.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 785 + }, + { + "path": "2094-142345-0012.flac", + "question": "Identify the portion of the clip where 'of' is most prominent.", + "choice_a": "[2.07, 2.48]", + "choice_b": "[9.52, 9.59]", + "choice_c": "[3.28, 14.0]", + "choice_d": "[8.15, 8.90]", + "answer_gt": "[9.52, 9.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 786 + }, + { + "path": "1998-15444-0000.flac", + "question": "In the audio track, when is 'case' first noticed?", + "choice_a": "[8.38, 8.88]", + "choice_b": "[1.14, 1.51]", + "choice_c": "[2.84, 2.98]", + "choice_d": "[0.98, 1.10]", + "answer_gt": "[1.14, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 787 + }, + { + "path": "4831-18525-0026.flac", + "question": "What is the duration of 'him' being mentioned in the conversation?", + "choice_a": "[3.30, 3.82]", + "choice_b": "[5.02, 5.14]", + "choice_c": "[2.88, 3.30]", + "choice_d": "[0.90, 1.34]", + "answer_gt": "[5.02, 5.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 788 + }, + { + "path": "8463-294828-0008.flac", + "question": "When does the audio feature the sound of 'what'?", + "choice_a": "[0.40, 0.69]", + "choice_b": "[0.89, 0.95]", + "choice_c": "[0.72, 0.89]", + "choice_d": "[0.21, 0.40]", + "answer_gt": "[0.72, 0.89]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 789 + }, + { + "path": "3576-138058-0032.flac", + "question": "Recognize the time when 'they' is mentioned by the speaker.", + "choice_a": "[3.04, 3.16]", + "choice_b": "[4.14, 4.43]", + "choice_c": "[3.16, 3.26]", + "choice_d": "[2.91, 3.04]", + "answer_gt": "[3.04, 3.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 790 + }, + { + "path": "5895-34629-0012.flac", + "question": "Find out the time when 'to' is immediately occured.", + "choice_a": "[1.31, 1.45]", + "choice_b": "[2.09, 2.49]", + "choice_c": "[2.49, 2.61]", + "choice_d": "[1.90, 2.09]", + "answer_gt": "[1.90, 2.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 791 + }, + { + "path": "6467-56885-0011.flac", + "question": "Note the time interval that includes the mention of 'the'.", + "choice_a": "[3.31, 3.72]", + "choice_b": "[0.14, 0.20]", + "choice_c": "[0.14, 0.20]", + "choice_d": "[1.15, 1.44]", + "answer_gt": "[0.14, 0.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 792 + }, + { + "path": "3081-166546-0024.flac", + "question": "Determine the time when 'a' happens within the clip.", + "choice_a": "[3.16, 3.25]", + "choice_b": "[3.56, 3.77]", + "choice_c": "[1.56, 1.62]", + "choice_d": "[2.75, 2.94]", + "answer_gt": "[1.56, 1.62]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 793 + }, + { + "path": "5442-41168-0009.flac", + "question": "Determine the precise moment 'to' in the audio's pace.", + "choice_a": "[2.82, 2.91]", + "choice_b": "[6.37, 7.14]", + "choice_c": "[1.63, 21.9]", + "choice_d": "[6.63, 27.1]", + "answer_gt": "[2.82, 2.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 794 + }, + { + "path": "7902-96591-0005.flac", + "question": "Pinpoint the exact moment 'the' is repeated in the audio.", + "choice_a": "[0.80, 1.00]", + "choice_b": "[0.60, 0.80]", + "choice_c": "[0.38, 0.60]", + "choice_d": "[1.55, 1.63]", + "answer_gt": "[1.55, 1.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 795 + }, + { + "path": "3660-6517-0010.flac", + "question": "At what time interval does 'three' occur in the audio clip?", + "choice_a": "[5.27, 5.67]", + "choice_b": "[1.34, 1.44]", + "choice_c": "[0.84, 1.03]", + "choice_d": "[6.71, 7.06]", + "answer_gt": "[0.84, 1.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 796 + }, + { + "path": "5694-64038-0017.flac", + "question": "Establish when 'emotions' is first mentioned in the clip.", + "choice_a": "[1.66, 12.1]", + "choice_b": "[2.54, 12.8]", + "choice_c": "[1.16, 11.6]", + "choice_d": "[4.80, 5.26]", + "answer_gt": "[4.80, 5.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 797 + }, + { + "path": "1255-74899-0002.flac", + "question": "Determine the point in the audio when 'the' is discussed.", + "choice_a": "[0.46, 0.53]", + "choice_b": "[1.24, 1.54]", + "choice_c": "[3.06, 3.31]", + "choice_d": "[1.68, 1.87]", + "answer_gt": "[0.46, 0.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 798 + }, + { + "path": "5639-40744-0013.flac", + "question": "Determine the point in the audio when 'a' is repeated.", + "choice_a": "[0.60, 0.85]", + "choice_b": "[4.18, 4.24]", + "choice_c": "[2.64, 2.83]", + "choice_d": "[1.56, 2.01]", + "answer_gt": "[4.18, 4.24]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 799 + }, + { + "path": "4350-10919-0031.flac", + "question": "Identify the moment 'to' is mentioned in the audio.", + "choice_a": "[0.85, 1.28]", + "choice_b": "[6.13, 6.40]", + "choice_c": "[2.71, 2.83]", + "choice_d": "[4.14, 4.28]", + "answer_gt": "[2.71, 2.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 800 + }, + { + "path": "61-70968-0052.flac", + "question": "Note the time interval that includes the mention of 'who'.", + "choice_a": "[0.33, 0.54]", + "choice_b": "[1.78, 1.86]", + "choice_c": "[1.26, 1.67]", + "choice_d": "[0.54, 0.66]", + "answer_gt": "[0.54, 0.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 801 + }, + { + "path": "4852-28312-0000.flac", + "question": "Pinpoint the exact moment 'mister' is repeated in the audio.", + "choice_a": "[2.66, 3.13]", + "choice_b": "[1.92, 2.20]", + "choice_c": "[1.92, 2.20]", + "choice_d": "[1.27, 1.35]", + "answer_gt": "[1.92, 2.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 802 + }, + { + "path": "3660-172182-0036.flac", + "question": "Determine the point in the audio when 'and' is repeated.", + "choice_a": "[2.23, 2.52]", + "choice_b": "[5.17, 5.46]", + "choice_c": "[0.37, 0.46]", + "choice_d": "[3.20, 3.26]", + "answer_gt": "[0.37, 0.46]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 803 + }, + { + "path": "5142-33396-0052.flac", + "question": "Highlight the time when 'hang' is mentioned.", + "choice_a": "[1.88, 2.00]", + "choice_b": "[0.97, 1.23]", + "choice_c": "[4.70, 4.93]", + "choice_d": "[0.37, 0.48]", + "answer_gt": "[4.70, 4.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 804 + }, + { + "path": "1580-141084-0005.flac", + "question": "Find out the time when 'said' is immediately occured.", + "choice_a": "[1.61, 1.82]", + "choice_b": "[0.50, 0.68]", + "choice_c": "[0.68, 0.82]", + "choice_d": "[1.04, 1.61]", + "answer_gt": "[1.61, 1.82]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 805 + }, + { + "path": "8842-302201-0012.flac", + "question": "Identify when 'in' is heard in the audio.", + "choice_a": "[1.38, 1.54]", + "choice_b": "[0.26, 0.39]", + "choice_c": "[1.68, 1.74]", + "choice_d": "[1.74, 2.08]", + "answer_gt": "[0.26, 0.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 806 + }, + { + "path": "260-123440-0006.flac", + "question": "Spot the exact moment 'wonder' appears in the audio.", + "choice_a": "[0.38, 0.64]", + "choice_b": "[0.29, 0.38]", + "choice_c": "[0.64, 0.76]", + "choice_d": "[0.82, 1.04]", + "answer_gt": "[0.38, 0.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 807 + }, + { + "path": "6070-86744-0025.flac", + "question": "Identify the moment 'he' is mentioned in the audio.", + "choice_a": "[0.82, 1.17]", + "choice_b": "[4.03, 4.46]", + "choice_c": "[4.57, 4.68]", + "choice_d": "[1.17, 1.28]", + "answer_gt": "[4.57, 4.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 808 + }, + { + "path": "2609-169640-0005.flac", + "question": "When in the audio does 'in' coincide with a specific event?", + "choice_a": "[0.37, 0.57]", + "choice_b": "[3.45, 3.94]", + "choice_c": "[2.63, 2.81]", + "choice_d": "[6.21, 6.37]", + "answer_gt": "[6.21, 6.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 809 + }, + { + "path": "7975-280085-0016.flac", + "question": "Establish when 'sheriff' is first mentioned in the clip.", + "choice_a": "[0.34, 0.66]", + "choice_b": "[3.71, 3.99]", + "choice_c": "[3.05, 3.39]", + "choice_d": "[4.49, 4.85]", + "answer_gt": "[0.34, 0.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 810 + }, + { + "path": "7641-96684-0001.flac", + "question": "Locate the segment where 'radiant' is faintly heard.", + "choice_a": "[2.95, 3.16]", + "choice_b": "[2.95, 3.16]", + "choice_c": "[4.53, 4.95]", + "choice_d": "[4.17, 4.28]", + "answer_gt": "[4.53, 4.95]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 811 + }, + { + "path": "2033-164916-0006.flac", + "question": "At what time interval does 'was' occur in the audio clip?", + "choice_a": "[0.53, 0.72]", + "choice_b": "[1.68, 1.97]", + "choice_c": "[0.85, 1.08]", + "choice_d": "[1.15, 1.68]", + "answer_gt": "[0.85, 1.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 812 + }, + { + "path": "6345-93306-0023.flac", + "question": "Spot the exact moment 'and' appears in the audio.", + "choice_a": "[0.47, 0.54]", + "choice_b": "[8.62, 8.91]", + "choice_c": "[0.87, 0.97]", + "choice_d": "[4.53, 4.87]", + "answer_gt": "[0.87, 0.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 813 + }, + { + "path": "700-122867-0008.flac", + "question": "Find the moment 'this' is clearly audible in the audio.", + "choice_a": "[0.24, 0.48]", + "choice_b": "[1.07, 1.43]", + "choice_c": "[2.75, 2.89]", + "choice_d": "[1.90, 2.07]", + "answer_gt": "[0.24, 0.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 814 + }, + { + "path": "5764-299665-0055.flac", + "question": "Find out the time when 'me' is immediately occured.", + "choice_a": "[5.33, 5.99]", + "choice_b": "[4.72, 5.15]", + "choice_c": "[0.86, 1.07]", + "choice_d": "[3.92, 4.05]", + "answer_gt": "[0.86, 1.07]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 815 + }, + { + "path": "1272-128104-0002.flac", + "question": "Recognize the time when 'of' is mentioned by the speaker.", + "choice_a": "[1.01, 1.32]", + "choice_b": "[4.17, 4.43]", + "choice_c": "[3.06, 3.22]", + "choice_d": "[5.88, 6.28]", + "answer_gt": "[3.06, 3.22]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 816 + }, + { + "path": "1650-167613-0004.flac", + "question": "Locate the part of the audio where 'understood' is expressed.", + "choice_a": "[1.50, 1.92]", + "choice_b": "[0.93, 1.50]", + "choice_c": "[5.60, 5.68]", + "choice_d": "[3.12, 3.73]", + "answer_gt": "[3.12, 3.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 817 + }, + { + "path": "2428-83705-0038.flac", + "question": "Identify the time segment when 'be' is mentioned in the audio.", + "choice_a": "[0.50, 0.89]", + "choice_b": "[4.38, 4.68]", + "choice_c": "[1.45, 1.52]", + "choice_d": "[2.82, 2.97]", + "answer_gt": "[4.38, 4.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 818 + }, + { + "path": "7105-2330-0023.flac", + "question": "Listen for when 'heaven's' is emphasized by the narrator.", + "choice_a": "[0.98, 1.31]", + "choice_b": "[0.36, 0.52]", + "choice_c": "[0.52, 0.98]", + "choice_d": "[0.36, 0.52]", + "answer_gt": "[0.52, 0.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 819 + }, + { + "path": "367-130732-0014.flac", + "question": "Listen for when 'knows' is emphasized by the narrator.", + "choice_a": "[1.27, 1.59]", + "choice_b": "[5.23, 5.60]", + "choice_c": "[0.33, 0.46]", + "choice_d": "[0.20, 0.33]", + "answer_gt": "[1.27, 1.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 820 + }, + { + "path": "2609-156975-0025.flac", + "question": "Highlight the time when 'moses' is mentioned.", + "choice_a": "[2.42, 2.85]", + "choice_b": "[0.90, 1.51]", + "choice_c": "[4.48, 4.71]", + "choice_d": "[1.92, 2.05]", + "answer_gt": "[0.90, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 821 + }, + { + "path": "4970-29095-0007.flac", + "question": "Identify the time segment when 'compelled' is mentioned in the audio.", + "choice_a": "[9.76, 9.98]", + "choice_b": "[0.56, 0.69]", + "choice_c": "[0.37, 10.7]", + "choice_d": "[3.59, 3.71]", + "answer_gt": "[0.37, 10.7]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 822 + }, + { + "path": "4572-64670-0023.flac", + "question": "At what time interval does 'and' occur in the audio clip?", + "choice_a": "[2.91, 3.07]", + "choice_b": "[2.11, 2.64]", + "choice_c": "[0.92, 1.01]", + "choice_d": "[1.01, 1.50]", + "answer_gt": "[0.92, 1.01]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 823 + }, + { + "path": "1630-73710-0020.flac", + "question": "When does 'near' occur in the audio?", + "choice_a": "[1.83, 2.13]", + "choice_b": "[0.60, 0.66]", + "choice_c": "[0.66, 1.44]", + "choice_d": "[2.13, 2.42]", + "answer_gt": "[2.13, 2.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 824 + }, + { + "path": "8254-84205-0012.flac", + "question": "Spot the exact moment 'start' appears in the audio.", + "choice_a": "[1.01, 1.37]", + "choice_b": "[1.88, 2.02]", + "choice_c": "[0.56, 0.80]", + "choice_d": "[0.88, 1.01]", + "answer_gt": "[1.01, 1.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 825 + }, + { + "path": "3752-4943-0018.flac", + "question": "Note the time interval that includes the mention of 'last'.", + "choice_a": "[0.75, 1.01]", + "choice_b": "[1.16, 1.42]", + "choice_c": "[3.17, 3.51]", + "choice_d": "[1.16, 1.42]", + "answer_gt": "[3.17, 3.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 826 + }, + { + "path": "6432-63722-0038.flac", + "question": "Recognize the time when 'free' is most clearly articulated.", + "choice_a": "[0.16, 0.30]", + "choice_b": "[2.97, 3.26]", + "choice_c": "[3.60, 3.99]", + "choice_d": "[2.72, 2.97]", + "answer_gt": "[2.97, 3.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 827 + }, + { + "path": "116-288045-0015.flac", + "question": "Pinpoint the passage where 'whispered' is heard in the background.", + "choice_a": "[0.72, 1.32]", + "choice_b": "[0.31, 0.63]", + "choice_c": "[0.63, 0.72]", + "choice_d": "[0.22, 0.31]", + "answer_gt": "[0.31, 0.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 828 + }, + { + "path": "8254-115543-0044.flac", + "question": "Identify the time segment when 'must' is mentioned in the audio.", + "choice_a": "[1.47, 1.85]", + "choice_b": "[1.23, 1.47]", + "choice_c": "[0.18, 0.44]", + "choice_d": "[0.44, 0.64]", + "answer_gt": "[0.18, 0.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 829 + }, + { + "path": "6467-97061-0014.flac", + "question": "Find the duration when 'day' occurs consecutively in the audio.", + "choice_a": "[3.97, 4.03]", + "choice_b": "[1.68, 1.79]", + "choice_c": "[6.91, 7.26]", + "choice_d": "[0.45, 0.75]", + "answer_gt": "[0.45, 0.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 830 + }, + { + "path": "8297-275154-0006.flac", + "question": "Pinpoint the exact moment 'over' appears.", + "choice_a": "[1.24, 1.40]", + "choice_b": "[4.22, 4.72]", + "choice_c": "[0.64, 10.9]", + "choice_d": "[1.42, 11.9]", + "answer_gt": "[0.64, 10.9]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 831 + }, + { + "path": "4198-12259-0033.flac", + "question": "Ascertain the exact time 'planets' is mentioned in the audio.", + "choice_a": "[0.37, 0.46]", + "choice_b": "[2.27, 2.45]", + "choice_c": "[0.46, 0.79]", + "choice_d": "[1.44, 2.09]", + "answer_gt": "[1.44, 2.09]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 832 + }, + { + "path": "3000-15664-0025.flac", + "question": "Identify when 'in' is heard in the audio.", + "choice_a": "[0.60, 1.16]", + "choice_b": "[1.34, 1.37]", + "choice_c": "[1.37, 1.77]", + "choice_d": "[3.86, 3.98]", + "answer_gt": "[3.86, 3.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 833 + }, + { + "path": "6829-68769-0027.flac", + "question": "Determine the precise moment 'check' in the audio's pace.", + "choice_a": "[2.06, 2.33]", + "choice_b": "[1.09, 1.35]", + "choice_c": "[1.50, 1.93]", + "choice_d": "[0.58, 0.81]", + "answer_gt": "[1.50, 1.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 834 + }, + { + "path": "2961-960-0015.flac", + "question": "What is the duration of 'plan' being mentioned in the conversation?", + "choice_a": "[0.56, 0.63]", + "choice_b": "[1.56, 2.03]", + "choice_c": "[4.44, 4.69]", + "choice_d": "[2.29, 3.06]", + "answer_gt": "[1.56, 2.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 835 + }, + { + "path": "1221-135766-0012.flac", + "question": "Note the time interval that includes the mention of 'incomprehensible'.", + "choice_a": "[1.20, 1.34]", + "choice_b": "[4.01, 4.42]", + "choice_c": "[8.18, 9.05]", + "choice_d": "[3.51, 14.6]", + "answer_gt": "[3.51, 14.6]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 836 + }, + { + "path": "8288-274162-0060.flac", + "question": "Listen for the occurrence of 'fact' in the recording.", + "choice_a": "[1.00, 1.08]", + "choice_b": "[3.39, 3.62]", + "choice_c": "[1.08, 1.52]", + "choice_d": "[2.08, 2.52]", + "answer_gt": "[1.08, 1.52]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 837 + }, + { + "path": "1988-24833-0022.flac", + "question": "Locate the segment where 'tom' appears most frequently.", + "choice_a": "[0.56, 0.61]", + "choice_b": "[0.88, 1.11]", + "choice_c": "[2.08, 2.31]", + "choice_d": "[1.78, 2.08]", + "answer_gt": "[0.88, 1.11]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 838 + }, + { + "path": "4507-16021-0038.flac", + "question": "What is the timestamp for the initial occurrence of 'this'?", + "choice_a": "[1.38, 1.55]", + "choice_b": "[0.68, 1.31]", + "choice_c": "[0.43, 0.68]", + "choice_d": "[1.61, 2.31]", + "answer_gt": "[0.43, 0.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 839 + }, + { + "path": "6313-66129-0028.flac", + "question": "Highlight when 'is' is spoken.", + "choice_a": "[0.54, 0.68]", + "choice_b": "[0.22, 0.54]", + "choice_c": "[0.68, 0.80]", + "choice_d": "[0.68, 0.80]", + "answer_gt": "[0.54, 0.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 840 + }, + { + "path": "1462-170142-0010.flac", + "question": "Ascertain the point at which 'she' enters the discussion.", + "choice_a": "[1.10, 1.21]", + "choice_b": "[3.06, 3.25]", + "choice_c": "[0.70, 1.10]", + "choice_d": "[1.21, 1.39]", + "answer_gt": "[3.06, 3.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 841 + }, + { + "path": "2033-164914-0004.flac", + "question": "Detect the point at which 'tell' is spoken.", + "choice_a": "[1.13, 1.56]", + "choice_b": "[2.86, 3.16]", + "choice_c": "[3.16, 3.32]", + "choice_d": "[0.49, 0.68]", + "answer_gt": "[2.86, 3.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 842 + }, + { + "path": "1701-141760-0013.flac", + "question": "Identify the moment 'you' is mentioned in the audio.", + "choice_a": "[1.25, 1.59]", + "choice_b": "[0.42, 0.61]", + "choice_c": "[0.61, 0.74]", + "choice_d": "[1.12, 1.25]", + "answer_gt": "[1.12, 1.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 843 + }, + { + "path": "6930-76324-0011.flac", + "question": "At what point in the audio is 'lovely' clearly audible?", + "choice_a": "[0.64, 0.79]", + "choice_b": "[7.94, 8.21]", + "choice_c": "[3.56, 3.88]", + "choice_d": "[6.93, 7.18]", + "answer_gt": "[3.56, 3.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 844 + }, + { + "path": "1701-141759-0024.flac", + "question": "Listen for the occurrence of 'been' in the recording.", + "choice_a": "[2.25, 2.38]", + "choice_b": "[1.29, 1.45]", + "choice_c": "[3.42, 3.75]", + "choice_d": "[1.17, 1.29]", + "answer_gt": "[1.29, 1.45]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 845 + }, + { + "path": "4507-16021-0059.flac", + "question": "At what time interval does 'therein' occur in the audio clip?", + "choice_a": "[3.56, 4.06]", + "choice_b": "[0.84, 1.22]", + "choice_c": "[1.22, 1.69]", + "choice_d": "[0.47, 0.84]", + "answer_gt": "[3.56, 4.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 846 + }, + { + "path": "3570-5694-0014.flac", + "question": "What is the timestamp for the initial occurrence of 'all'?", + "choice_a": "[2.11, 2.47]", + "choice_b": "[7.40, 7.68]", + "choice_c": "[3.95, 4.11]", + "choice_d": "[3.10, 3.38]", + "answer_gt": "[7.40, 7.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 847 + }, + { + "path": "1255-90413-0008.flac", + "question": "Identify the moment 'pull' is mentioned in the audio.", + "choice_a": "[0.55, 0.67]", + "choice_b": "[1.97, 2.30]", + "choice_c": "[4.41, 4.89]", + "choice_d": "[5.27, 5.37]", + "answer_gt": "[1.97, 2.30]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 848 + }, + { + "path": "700-122866-0034.flac", + "question": "When does the audio feature the sound of 'i'm'?", + "choice_a": "[0.92, 1.29]", + "choice_b": "[0.38, 0.79]", + "choice_c": "[0.25, 0.38]", + "choice_d": "[1.29, 1.37]", + "answer_gt": "[0.25, 0.38]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 849 + }, + { + "path": "6829-68769-0053.flac", + "question": "Find the moment 'save' is clearly audible in the audio.", + "choice_a": "[2.04, 2.28]", + "choice_b": "[4.01, 4.09]", + "choice_c": "[1.40, 1.78]", + "choice_d": "[1.23, 1.40]", + "answer_gt": "[1.40, 1.78]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 850 + }, + { + "path": "251-136532-0013.flac", + "question": "When does the audio feature the sound of 'notice'?", + "choice_a": "[1.84, 1.99]", + "choice_b": "[1.99, 2.07]", + "choice_c": "[2.98, 3.35]", + "choice_d": "[3.35, 3.52]", + "answer_gt": "[2.98, 3.35]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 851 + }, + { + "path": "5543-27761-0004.flac", + "question": "Determine the precise moment 'mother' in the audio's pace.", + "choice_a": "[0.52, 0.61]", + "choice_b": "[3.37, 3.45]", + "choice_c": "[3.06, 3.37]", + "choice_d": "[3.45, 3.93]", + "answer_gt": "[3.06, 3.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 852 + }, + { + "path": "1255-90407-0010.flac", + "question": "Listen to this audio clip: when does 'have' occur?", + "choice_a": "[0.66, 0.79]", + "choice_b": "[0.79, 0.98]", + "choice_c": "[0.79, 0.98]", + "choice_d": "[0.79, 0.98]", + "answer_gt": "[0.66, 0.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 853 + }, + { + "path": "7176-92135-0001.flac", + "question": "When does the speaker mention 'london'?", + "choice_a": "[2.17, 2.52]", + "choice_b": "[4.57, 4.81]", + "choice_c": "[2.78, 3.06]", + "choice_d": "[5.27, 5.74]", + "answer_gt": "[2.78, 3.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 854 + }, + { + "path": "4350-10919-0033.flac", + "question": "Pinpoint the passage where 'kitty' is heard in the background.", + "choice_a": "[2.38, 2.45]", + "choice_b": "[3.41, 3.59]", + "choice_c": "[2.27, 2.38]", + "choice_d": "[0.49, 0.72]", + "answer_gt": "[3.41, 3.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 855 + }, + { + "path": "7850-286674-0016.flac", + "question": "Locate the segment where 'to' appears most frequently.", + "choice_a": "[0.82, 0.93]", + "choice_b": "[6.52, 6.89]", + "choice_c": "[9.65, 9.88]", + "choice_d": "[1.92, 2.72]", + "answer_gt": "[0.82, 0.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 856 + }, + { + "path": "6070-86744-0016.flac", + "question": "Listen to this audio clip; identify the time segment when 'him' is spoken.", + "choice_a": "[0.74, 0.87]", + "choice_b": "[1.42, 1.56]", + "choice_c": "[1.56, 2.12]", + "choice_d": "[2.12, 2.23]", + "answer_gt": "[1.42, 1.56]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 857 + }, + { + "path": "8254-84205-0047.flac", + "question": "When does 'yes' occur in the audio?", + "choice_a": "[0.86, 0.95]", + "choice_b": "[0.67, 0.86]", + "choice_c": "[2.17, 2.37]", + "choice_d": "[0.42, 0.46]", + "answer_gt": "[2.17, 2.37]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 858 + }, + { + "path": "5338-24640-0003.flac", + "question": "Determine the time when 'to' happens within the clip.", + "choice_a": "[0.79, 1.13]", + "choice_b": "[4.39, 4.50]", + "choice_c": "[2.34, 2.64]", + "choice_d": "[0.57, 0.79]", + "answer_gt": "[4.39, 4.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 859 + }, + { + "path": "6267-65525-0018.flac", + "question": "When in the audio does 'answered' coincide with a specific event?", + "choice_a": "[0.99, 1.32]", + "choice_b": "[2.10, 2.64]", + "choice_c": "[1.40, 1.80]", + "choice_d": "[1.32, 1.40]", + "answer_gt": "[2.10, 2.64]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 860 + }, + { + "path": "1701-141759-0016.flac", + "question": "Discover when 'speaking' is introduced in the conversation.", + "choice_a": "[4.18, 4.28]", + "choice_b": "[1.34, 1.47]", + "choice_c": "[1.47, 1.88]", + "choice_d": "[0.32, 0.48]", + "answer_gt": "[1.47, 1.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 861 + }, + { + "path": "4515-11057-0007.flac", + "question": "Locate the segment where 'me' appears most frequently.", + "choice_a": "[1.10, 1.51]", + "choice_b": "[2.17, 2.42]", + "choice_c": "[0.99, 1.10]", + "choice_d": "[2.54, 2.79]", + "answer_gt": "[0.99, 1.10]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 862 + }, + { + "path": "5683-32866-0003.flac", + "question": "At what interval does 'formed' appear in the audio?", + "choice_a": "[1.07, 1.19]", + "choice_b": "[1.29, 1.59]", + "choice_c": "[0.39, 0.44]", + "choice_d": "[1.82, 2.16]", + "answer_gt": "[1.29, 1.59]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 863 + }, + { + "path": "5694-64038-0019.flac", + "question": "Listen for the precise interval when 'to' is dominant.", + "choice_a": "[0.66, 0.72]", + "choice_b": "[2.62, 2.84]", + "choice_c": "[5.00, 5.14]", + "choice_d": "[2.84, 3.34]", + "answer_gt": "[0.66, 0.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 864 + }, + { + "path": "174-50561-0015.flac", + "question": "At what time interval does 'now' occur in the audio clip?", + "choice_a": "[7.92, 8.34]", + "choice_b": "[6.17, 6.28]", + "choice_c": "[0.53, 0.80]", + "choice_d": "[4.49, 4.82]", + "answer_gt": "[0.53, 0.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 865 + }, + { + "path": "1919-142785-0012.flac", + "question": "Find the duration when 'clean' occurs consecutively in the audio.", + "choice_a": "[2.91, 3.25]", + "choice_b": "[4.38, 4.52]", + "choice_c": "[1.97, 2.04]", + "choice_d": "[2.04, 2.37]", + "answer_gt": "[2.91, 3.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 866 + }, + { + "path": "3080-5032-0016.flac", + "question": "When does the speaker mention 'altered'?", + "choice_a": "[3.18, 3.71]", + "choice_b": "[1.54, 2.08]", + "choice_c": "[0.45, 0.57]", + "choice_d": "[2.45, 2.82]", + "answer_gt": "[1.54, 2.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 867 + }, + { + "path": "8254-84205-0063.flac", + "question": "When does 'but' occur in the audio?", + "choice_a": "[2.43, 2.51]", + "choice_b": "[0.33, 0.44]", + "choice_c": "[2.92, 3.01]", + "choice_d": "[0.44, 0.75]", + "answer_gt": "[0.33, 0.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 868 + }, + { + "path": "4198-61336-0025.flac", + "question": "Recognize the time when 'the' is mentioned by the speaker.", + "choice_a": "[1.50, 2.27]", + "choice_b": "[1.27, 1.42]", + "choice_c": "[0.66, 1.24]", + "choice_d": "[0.56, 0.66]", + "answer_gt": "[0.56, 0.66]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 869 + }, + { + "path": "672-122797-0043.flac", + "question": "Listen for the occurrence of 'meaning' in the recording.", + "choice_a": "[0.60, 0.93]", + "choice_b": "[1.08, 1.58]", + "choice_c": "[0.93, 1.08]", + "choice_d": "[1.80, 2.39]", + "answer_gt": "[1.08, 1.58]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 870 + }, + { + "path": "8188-269288-0034.flac", + "question": "Note the time interval that includes the mention of 'in'.", + "choice_a": "[1.65, 2.00]", + "choice_b": "[2.77, 3.21]", + "choice_c": "[0.57, 0.85]", + "choice_d": "[5.58, 5.69]", + "answer_gt": "[5.58, 5.69]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 871 + }, + { + "path": "5484-24318-0025.flac", + "question": "Locate the segment where 'i' is faintly heard.", + "choice_a": "[3.65, 3.79]", + "choice_b": "[0.44, 0.77]", + "choice_c": "[3.79, 3.98]", + "choice_d": "[0.77, 1.32]", + "answer_gt": "[0.44, 0.77]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 872 + }, + { + "path": "5142-36377-0014.flac", + "question": "Find out the time when 'too' is immediately occured.", + "choice_a": "[1.06, 1.19]", + "choice_b": "[7.35, 7.82]", + "choice_c": "[4.35, 14.8]", + "choice_d": "[3.44, 3.88]", + "answer_gt": "[3.44, 3.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 873 + }, + { + "path": "4294-35475-0014.flac", + "question": "Pinpoint the passage where 'its' is heard in the background.", + "choice_a": "[0.27, 0.52]", + "choice_b": "[4.95, 5.13]", + "choice_c": "[3.22, 3.53]", + "choice_d": "[0.98, 1.05]", + "answer_gt": "[4.95, 5.13]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 874 + }, + { + "path": "4323-55228-0039.flac", + "question": "Determine the point in the audio when 'i' is repeated.", + "choice_a": "[2.26, 2.39]", + "choice_b": "[0.17, 0.26]", + "choice_c": "[0.17, 0.26]", + "choice_d": "[2.39, 2.70]", + "answer_gt": "[0.17, 0.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 875 + }, + { + "path": "5895-34622-0008.flac", + "question": "When in the audio does 'improved' coincide with a specific event?", + "choice_a": "[6.01, 6.50]", + "choice_b": "[6.63, 6.71]", + "choice_c": "[0.80, 0.92]", + "choice_d": "[0.44, 0.63]", + "answer_gt": "[6.01, 6.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 876 + }, + { + "path": "1221-135767-0015.flac", + "question": "Pinpoint the passage where 'worship' is heard in the background.", + "choice_a": "[0.50, 0.69]", + "choice_b": "[1.51, 1.86]", + "choice_c": "[1.25, 1.51]", + "choice_d": "[0.35, 0.50]", + "answer_gt": "[1.51, 1.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 877 + }, + { + "path": "533-131564-0005.flac", + "question": "When does 'bit' occur in the audio?", + "choice_a": "[1.01, 1.07]", + "choice_b": "[4.69, 5.17]", + "choice_c": "[1.07, 1.29]", + "choice_d": "[3.27, 3.69]", + "answer_gt": "[1.07, 1.29]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 878 + }, + { + "path": "777-126732-0064.flac", + "question": "What is the duration of 'pause' being mentioned in the conversation?", + "choice_a": "[3.50, 3.93]", + "choice_b": "[2.53, 2.64]", + "choice_c": "[2.64, 3.10]", + "choice_d": "[1.40, 1.64]", + "answer_gt": "[3.50, 3.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 879 + }, + { + "path": "6599-38590-0006.flac", + "question": "Determine the precise moment 'give' in the audio's pace.", + "choice_a": "[1.86, 2.08]", + "choice_b": "[0.38, 0.53]", + "choice_c": "[0.64, 0.95]", + "choice_d": "[0.98, 1.13]", + "answer_gt": "[0.38, 0.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 880 + }, + { + "path": "5442-41169-0021.flac", + "question": "When is 'but' first heard in the recording?", + "choice_a": "[7.85, 8.41]", + "choice_b": "[7.46, 7.68]", + "choice_c": "[1.50, 1.62]", + "choice_d": "[4.00, 4.40]", + "answer_gt": "[7.46, 7.68]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 881 + }, + { + "path": "1701-141759-0025.flac", + "question": "In the audio track, when is 'an' first noticed?", + "choice_a": "[5.94, 6.10]", + "choice_b": "[2.68, 2.81]", + "choice_c": "[9.93, 10.2]", + "choice_d": "[9.67, 9.93]", + "answer_gt": "[2.68, 2.81]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 882 + }, + { + "path": "3170-137482-0013.flac", + "question": "Listen to this audio clip: when does 'where' occur?", + "choice_a": "[1.03, 1.19]", + "choice_b": "[0.88, 1.03]", + "choice_c": "[0.88, 1.03]", + "choice_d": "[0.52, 0.88]", + "answer_gt": "[0.52, 0.88]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 883 + }, + { + "path": "4294-9934-0021.flac", + "question": "Highlight the time when 'twenty' is mentioned.", + "choice_a": "[1.48, 1.86]", + "choice_b": "[0.91, 1.31]", + "choice_c": "[1.40, 1.48]", + "choice_d": "[0.63, 0.91]", + "answer_gt": "[0.63, 0.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 884 + }, + { + "path": "1701-141760-0005.flac", + "question": "Listen to this audio clip: when does 'to' occur?", + "choice_a": "[0.85, 1.03]", + "choice_b": "[0.51, 10.8]", + "choice_c": "[7.16, 7.20]", + "choice_d": "[0.38, 10.5]", + "answer_gt": "[0.38, 10.5]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 885 + }, + { + "path": "7850-281318-0019.flac", + "question": "Identify the moment 'your' is mentioned in the audio.", + "choice_a": "[4.30, 4.52]", + "choice_b": "[3.61, 3.70]", + "choice_c": "[2.67, 2.91]", + "choice_d": "[1.91, 2.25]", + "answer_gt": "[3.61, 3.70]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 886 + }, + { + "path": "3729-6852-0030.flac", + "question": "What is the timestamp for the initial occurrence of 'than'?", + "choice_a": "[1.12, 1.47]", + "choice_b": "[1.65, 2.00]", + "choice_c": "[0.71, 0.87]", + "choice_d": "[1.47, 1.65]", + "answer_gt": "[1.47, 1.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 887 + }, + { + "path": "7601-101622-0002.flac", + "question": "Determine the time when 'along' happens within the clip.", + "choice_a": "[6.41, 6.47]", + "choice_b": "[0.96, 1.36]", + "choice_c": "[9.50, 9.87]", + "choice_d": "[1.60, 1.84]", + "answer_gt": "[9.50, 9.87]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 888 + }, + { + "path": "6313-66125-0010.flac", + "question": "Listen for the occurrence of 'both' in the recording.", + "choice_a": "[2.71, 2.89]", + "choice_b": "[1.97, 2.07]", + "choice_c": "[2.07, 2.16]", + "choice_d": "[0.60, 0.79]", + "answer_gt": "[0.60, 0.79]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 889 + }, + { + "path": "6313-76958-0008.flac", + "question": "Identify the moment 'curley' is mentioned in the audio.", + "choice_a": "[0.45, 0.87]", + "choice_b": "[0.87, 1.38]", + "choice_c": "[0.87, 1.38]", + "choice_d": "[1.38, 1.75]", + "answer_gt": "[1.38, 1.75]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 890 + }, + { + "path": "121-121726-0012.flac", + "question": "Find the duration when 'bond' occurs consecutively in the audio.", + "choice_a": "[0.94, 1.41]", + "choice_b": "[1.68, 2.07]", + "choice_c": "[0.20, 0.71]", + "choice_d": "[2.07, 2.61]", + "answer_gt": "[2.07, 2.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 891 + }, + { + "path": "5142-36377-0021.flac", + "question": "When does the speaker mention 'sat'?", + "choice_a": "[7.61, 17.8]", + "choice_b": "[3.06, 3.22]", + "choice_c": "[7.11, 7.43]", + "choice_d": "[6.14, 6.28]", + "answer_gt": "[7.61, 17.8]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 892 + }, + { + "path": "700-122867-0040.flac", + "question": "When in the audio does 'is' coincide with a specific event?", + "choice_a": "[0.69, 0.86]", + "choice_b": "[0.86, 1.01]", + "choice_c": "[0.53, 0.69]", + "choice_d": "[0.53, 0.69]", + "answer_gt": "[0.86, 1.01]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 893 + }, + { + "path": "700-122867-0038.flac", + "question": "Determine the time when 'my' happens within the clip.", + "choice_a": "[5.46, 5.57]", + "choice_b": "[3.32, 3.62]", + "choice_c": "[1.33, 1.48]", + "choice_d": "[3.27, 3.32]", + "answer_gt": "[1.33, 1.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 894 + }, + { + "path": "116-288048-0016.flac", + "question": "At what point in the audio is 'saw' clearly audible?", + "choice_a": "[0.97, 1.16]", + "choice_b": "[1.16, 1.22]", + "choice_c": "[0.77, 0.97]", + "choice_d": "[1.22, 1.52]", + "answer_gt": "[1.22, 1.52]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 895 + }, + { + "path": "672-122797-0006.flac", + "question": "Ascertain the point at which 'older' enters the discussion.", + "choice_a": "[6.84, 7.34]", + "choice_b": "[3.58, 4.48]", + "choice_c": "[0.94, 11.8]", + "choice_d": "[1.91, 12.4]", + "answer_gt": "[3.58, 4.48]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 896 + }, + { + "path": "4323-55228-0009.flac", + "question": "Locate the part of the audio where 'before' is expressed.", + "choice_a": "[4.05, 4.28]", + "choice_b": "[4.67, 4.96]", + "choice_c": "[0.79, 1.12]", + "choice_d": "[1.28, 1.53]", + "answer_gt": "[1.28, 1.53]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 897 + }, + { + "path": "5105-28240-0016.flac", + "question": "When is 'the' first mentioned in the clip?", + "choice_a": "[0.93, 1.13]", + "choice_b": "[2.29, 2.93]", + "choice_c": "[1.94, 2.26]", + "choice_d": "[1.88, 1.94]", + "answer_gt": "[1.88, 1.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 898 + }, + { + "path": "367-293981-0006.flac", + "question": "Recognize the segment where 'enchantment' is spoken by the speaker.", + "choice_a": "[4.89, 5.21]", + "choice_b": "[2.20, 2.61]", + "choice_c": "[4.42, 4.78]", + "choice_d": "[3.37, 3.92]", + "answer_gt": "[3.37, 3.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 899 + }, + { + "path": "1462-170142-0026.flac", + "question": "Find out the time when 'hand' is immediately occured.", + "choice_a": "[1.40, 1.82]", + "choice_b": "[4.08, 4.18]", + "choice_c": "[1.14, 1.40]", + "choice_d": "[0.42, 0.60]", + "answer_gt": "[1.14, 1.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 900 + }, + { + "path": "1221-135767-0016.flac", + "question": "At what time interval does 'and' occur in the audio clip?", + "choice_a": "[5.58, 5.72]", + "choice_b": "[2.87, 3.02]", + "choice_c": "[7.01, 7.30]", + "choice_d": "[2.87, 3.02]", + "answer_gt": "[5.58, 5.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 901 + }, + { + "path": "7850-73752-0002.flac", + "question": "Determine the precise moment 'to' in the audio's pace.", + "choice_a": "[2.75, 3.39]", + "choice_b": "[2.23, 2.32]", + "choice_c": "[1.20, 1.44]", + "choice_d": "[5.77, 6.53]", + "answer_gt": "[2.23, 2.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 902 + }, + { + "path": "7105-2330-0018.flac", + "question": "Ascertain the exact time 'a' is mentioned in the audio.", + "choice_a": "[2.20, 2.59]", + "choice_b": "[4.84, 5.04]", + "choice_c": "[4.47, 4.66]", + "choice_d": "[2.59, 2.65]", + "answer_gt": "[2.59, 2.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 903 + }, + { + "path": "1580-141083-0026.flac", + "question": "Recognize the time when 'by' is mentioned by the speaker.", + "choice_a": "[0.99, 1.07]", + "choice_b": "[1.50, 1.57]", + "choice_c": "[3.36, 3.50]", + "choice_d": "[0.71, 0.99]", + "answer_gt": "[3.36, 3.50]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 904 + }, + { + "path": "2428-83705-0001.flac", + "question": "At what point in the audio is 'be' clearly audible?", + "choice_a": "[1.42, 1.54]", + "choice_b": "[4.77, 4.93]", + "choice_c": "[2.16, 2.29]", + "choice_d": "[2.29, 2.84]", + "answer_gt": "[4.77, 4.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 905 + }, + { + "path": "8188-269290-0015.flac", + "question": "Note the time interval that includes the mention of 'you'.", + "choice_a": "[1.21, 1.31]", + "choice_b": "[2.53, 2.86]", + "choice_c": "[1.31, 1.55]", + "choice_d": "[2.38, 2.47]", + "answer_gt": "[1.21, 1.31]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 906 + }, + { + "path": "4515-11057-0080.flac", + "question": "Locate the segment where 'advise' appears most frequently.", + "choice_a": "[0.76, 1.11]", + "choice_b": "[1.61, 1.97]", + "choice_c": "[2.89, 3.00]", + "choice_d": "[4.28, 4.84]", + "answer_gt": "[4.28, 4.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 907 + }, + { + "path": "7105-2340-0004.flac", + "question": "Note the time interval that includes the mention of 'cousin'.", + "choice_a": "[7.40, 7.91]", + "choice_b": "[6.65, 6.96]", + "choice_c": "[1.09, 11.4]", + "choice_d": "[0.89, 1.65]", + "answer_gt": "[7.40, 7.91]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 908 + }, + { + "path": "6345-93306-0009.flac", + "question": "Pinpoint the exact moment 'cut' is repeated in the audio.", + "choice_a": "[0.89, 1.14]", + "choice_b": "[1.51, 1.77]", + "choice_c": "[0.28, 0.51]", + "choice_d": "[1.40, 1.51]", + "answer_gt": "[0.89, 1.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 909 + }, + { + "path": "3005-163389-0015.flac", + "question": "Pinpoint the exact moment 'of' appears.", + "choice_a": "[3.95, 4.25]", + "choice_b": "[1.39, 1.72]", + "choice_c": "[0.68, 0.97]", + "choice_d": "[3.31, 3.40]", + "answer_gt": "[3.31, 3.40]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 910 + }, + { + "path": "3005-163389-0004.flac", + "question": "When is 'man's' first mentioned in the clip?", + "choice_a": "[1.09, 1.51]", + "choice_b": "[0.64, 1.09]", + "choice_c": "[0.52, 0.64]", + "choice_d": "[0.52, 0.64]", + "answer_gt": "[1.09, 1.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 911 + }, + { + "path": "7850-73752-0019.flac", + "question": "Highlight the time when 'miss' is mentioned.", + "choice_a": "[1.66, 2.13]", + "choice_b": "[1.00, 1.53]", + "choice_c": "[0.56, 0.69]", + "choice_d": "[2.13, 2.34]", + "answer_gt": "[2.13, 2.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 912 + }, + { + "path": "3000-15664-0022.flac", + "question": "Highlight when 'here' is spoken.", + "choice_a": "[0.27, 0.41]", + "choice_b": "[0.99, 1.08]", + "choice_c": "[2.69, 2.91]", + "choice_d": "[0.41, 0.57]", + "answer_gt": "[0.27, 0.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 913 + }, + { + "path": "3528-168656-0006.flac", + "question": "At what time interval does 'not' occur in the audio clip?", + "choice_a": "[0.77, 0.97]", + "choice_b": "[1.72, 1.82]", + "choice_c": "[0.39, 0.61]", + "choice_d": "[0.61, 0.77]", + "answer_gt": "[0.77, 0.97]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 914 + }, + { + "path": "6345-93306-0025.flac", + "question": "Determine the precise moment 'in' in the audio's pace.", + "choice_a": "[3.12, 3.42]", + "choice_b": "[0.27, 10.5]", + "choice_c": "[0.02, 10.2]", + "choice_d": "[3.99, 4.12]", + "answer_gt": "[3.99, 4.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 915 + }, + { + "path": "6841-88294-0003.flac", + "question": "In the audio track, when is 'had' first noticed?", + "choice_a": "[1.46, 1.60]", + "choice_b": "[2.58, 2.82]", + "choice_c": "[3.77, 3.83]", + "choice_d": "[0.46, 0.92]", + "answer_gt": "[1.46, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 916 + }, + { + "path": "6313-66129-0007.flac", + "question": "At what point in the audio is 'that' clearly audible?", + "choice_a": "[0.77, 0.92]", + "choice_b": "[0.51, 0.77]", + "choice_c": "[0.34, 0.51]", + "choice_d": "[1.25, 1.34]", + "answer_gt": "[1.25, 1.34]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 917 + }, + { + "path": "1988-147956-0013.flac", + "question": "Listen for the occurrence of 'the' in the recording.", + "choice_a": "[1.01, 1.20]", + "choice_b": "[0.63, 0.76]", + "choice_c": "[2.26, 2.32]", + "choice_d": "[1.94, 2.00]", + "answer_gt": "[1.94, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 918 + }, + { + "path": "8254-84205-0011.flac", + "question": "Listen for when 'in' is emphasized by the narrator.", + "choice_a": "[3.93, 4.03]", + "choice_b": "[3.47, 3.93]", + "choice_c": "[3.35, 3.47]", + "choice_d": "[1.42, 1.83]", + "answer_gt": "[3.93, 4.03]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 919 + }, + { + "path": "3080-5040-0029.flac", + "question": "Listen to this audio clip: when does 'is' occur?", + "choice_a": "[0.09, 10.3]", + "choice_b": "[3.62, 3.79]", + "choice_c": "[0.71, 10.8]", + "choice_d": "[1.23, 1.61]", + "answer_gt": "[0.71, 10.8]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 920 + }, + { + "path": "5764-299665-0037.flac", + "question": "When in the audio does 'human' coincide with a specific event?", + "choice_a": "[1.54, 1.98]", + "choice_b": "[5.01, 5.39]", + "choice_c": "[2.47, 2.57]", + "choice_d": "[2.60, 3.21]", + "answer_gt": "[1.54, 1.98]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 921 + }, + { + "path": "8455-210777-0049.flac", + "question": "Locate the segment where 'lieutenant' is faintly heard.", + "choice_a": "[1.12, 1.87]", + "choice_b": "[0.50, 1.12]", + "choice_c": "[2.18, 2.50]", + "choice_d": "[2.11, 2.18]", + "answer_gt": "[0.50, 1.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 922 + }, + { + "path": "1221-135767-0024.flac", + "question": "Determine the precise moment 'the' in the audio's pace.", + "choice_a": "[0.48, 0.90]", + "choice_b": "[3.20, 3.23]", + "choice_c": "[2.43, 2.53]", + "choice_d": "[1.29, 1.36]", + "answer_gt": "[1.29, 1.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 923 + }, + { + "path": "8455-210777-0061.flac", + "question": "Discover when 'is' is introduced in the conversation.", + "choice_a": "[8.53, 8.72]", + "choice_b": "[0.41, 0.52]", + "choice_c": "[3.16, 3.78]", + "choice_d": "[6.40, 6.55]", + "answer_gt": "[8.53, 8.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 924 + }, + { + "path": "7697-105817-0005.flac", + "question": "Recognize the time when 'through' is most clearly articulated.", + "choice_a": "[1.28, 1.44]", + "choice_b": "[1.44, 1.55]", + "choice_c": "[0.53, 0.65]", + "choice_d": "[0.65, 0.78]", + "answer_gt": "[1.28, 1.44]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 925 + }, + { + "path": "3997-180297-0004.flac", + "question": "Recognize the segment where 'what' is spoken by the speaker.", + "choice_a": "[3.10, 3.47]", + "choice_b": "[1.93, 2.47]", + "choice_c": "[3.63, 3.85]", + "choice_d": "[1.81, 1.93]", + "answer_gt": "[3.63, 3.85]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 926 + }, + { + "path": "1255-138279-0017.flac", + "question": "Locate the part of the audio where 'which' is expressed.", + "choice_a": "[0.71, 0.97]", + "choice_b": "[1.45, 1.91]", + "choice_c": "[0.59, 0.71]", + "choice_d": "[0.97, 1.12]", + "answer_gt": "[0.97, 1.12]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 927 + }, + { + "path": "5895-34629-0007.flac", + "question": "Determine the point in the audio when 'the' is discussed.", + "choice_a": "[3.52, 3.94]", + "choice_b": "[0.55, 0.61]", + "choice_c": "[6.32, 6.48]", + "choice_d": "[0.55, 0.61]", + "answer_gt": "[0.55, 0.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 928 + }, + { + "path": "5442-41168-0004.flac", + "question": "Identify the portion of the clip where 'for' is most prominent.", + "choice_a": "[0.43, 0.73]", + "choice_b": "[1.23, 1.41]", + "choice_c": "[0.26, 0.43]", + "choice_d": "[0.73, 0.85]", + "answer_gt": "[1.23, 1.41]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 929 + }, + { + "path": "6829-68769-0024.flac", + "question": "When in the audio does 'rogers' coincide with a specific event?", + "choice_a": "[2.19, 2.31]", + "choice_b": "[2.71, 3.01]", + "choice_c": "[3.34, 3.74]", + "choice_d": "[1.81, 1.95]", + "answer_gt": "[3.34, 3.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 930 + }, + { + "path": "1272-135031-0002.flac", + "question": "Pinpoint the passage where 'the' is heard in the background.", + "choice_a": "[7.96, 8.04]", + "choice_b": "[3.62, 3.96]", + "choice_c": "[1.29, 1.87]", + "choice_d": "[6.45, 6.51]", + "answer_gt": "[7.96, 8.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 931 + }, + { + "path": "4198-12259-0042.flac", + "question": "Find out the time when 'thirsty' is immediately occured.", + "choice_a": "[4.18, 4.61]", + "choice_b": "[2.40, 2.59]", + "choice_c": "[2.70, 3.18]", + "choice_d": "[2.59, 2.70]", + "answer_gt": "[4.18, 4.61]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 932 + }, + { + "path": "7105-2330-0037.flac", + "question": "Establish when 'song' is first mentioned in the clip.", + "choice_a": "[2.84, 3.30]", + "choice_b": "[1.24, 1.67]", + "choice_c": "[4.07, 4.41]", + "choice_d": "[3.39, 4.07]", + "answer_gt": "[1.24, 1.67]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 933 + }, + { + "path": "1255-90407-0022.flac", + "question": "Recognize the segment where 'and' is spoken by the speaker.", + "choice_a": "[5.02, 5.53]", + "choice_b": "[2.91, 3.05]", + "choice_c": "[4.16, 4.26]", + "choice_d": "[4.35, 4.72]", + "answer_gt": "[4.16, 4.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 934 + }, + { + "path": "5895-34629-0019.flac", + "question": "Ascertain the point at which 'london' enters the discussion.", + "choice_a": "[2.48, 2.86]", + "choice_b": "[0.42, 0.48]", + "choice_c": "[0.48, 0.88]", + "choice_d": "[0.88, 1.03]", + "answer_gt": "[2.48, 2.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 935 + }, + { + "path": "1255-74899-0009.flac", + "question": "Find out the time when 'from' is immediately occured.", + "choice_a": "[0.59, 0.66]", + "choice_b": "[1.45, 1.63]", + "choice_c": "[1.07, 1.45]", + "choice_d": "[0.66, 1.07]", + "answer_gt": "[1.45, 1.63]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 936 + }, + { + "path": "6599-38591-0003.flac", + "question": "Find the moment 'he' is clearly audible in the audio.", + "choice_a": "[0.39, 0.49]", + "choice_b": "[0.86, 0.92]", + "choice_c": "[1.35, 1.65]", + "choice_d": "[0.92, 1.35]", + "answer_gt": "[0.86, 0.92]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 937 + }, + { + "path": "3005-163391-0011.flac", + "question": "What is the timestamp for the initial occurrence of 'was'?", + "choice_a": "[7.58, 7.65]", + "choice_b": "[1.18, 1.32]", + "choice_c": "[3.16, 13.4]", + "choice_d": "[8.86, 9.20]", + "answer_gt": "[1.18, 1.32]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 938 + }, + { + "path": "7176-92135-0036.flac", + "question": "Determine the precise moment 'can' in the audio's pace.", + "choice_a": "[0.51, 0.99]", + "choice_b": "[0.20, 0.48]", + "choice_c": "[4.01, 4.10]", + "choice_d": "[5.56, 5.76]", + "answer_gt": "[5.56, 5.76]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 939 + }, + { + "path": "4446-2273-0032.flac", + "question": "Listen for the precise interval when 'said' is dominant.", + "choice_a": "[3.04, 3.42]", + "choice_b": "[6.58, 6.63]", + "choice_c": "[2.05, 2.12]", + "choice_d": "[1.82, 2.05]", + "answer_gt": "[3.04, 3.42]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 940 + }, + { + "path": "6455-67804-0021.flac", + "question": "Note the time when 'barold' is first mentioned in the audio.", + "choice_a": "[0.36, 0.60]", + "choice_b": "[4.43, 4.62]", + "choice_c": "[2.51, 2.80]", + "choice_d": "[0.94, 1.29]", + "answer_gt": "[2.51, 2.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 941 + }, + { + "path": "1988-24833-0017.flac", + "question": "Recognize the time when 'minutes' is mentioned by the speaker.", + "choice_a": "[3.08, 3.36]", + "choice_b": "[2.09, 2.55]", + "choice_c": "[0.57, 0.74]", + "choice_d": "[4.11, 4.34]", + "answer_gt": "[2.09, 2.55]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 942 + }, + { + "path": "6313-66129-0023.flac", + "question": "At what interval does 'into' appear in the audio?", + "choice_a": "[1.96, 2.06]", + "choice_b": "[0.76, 0.93]", + "choice_c": "[1.31, 1.38]", + "choice_d": "[2.79, 2.84]", + "answer_gt": "[0.76, 0.93]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 943 + }, + { + "path": "6930-81414-0017.flac", + "question": "Listen for when 'it' is emphasized by the narrator.", + "choice_a": "[1.38, 1.69]", + "choice_b": "[0.64, 0.86]", + "choice_c": "[0.19, 0.36]", + "choice_d": "[1.02, 1.18]", + "answer_gt": "[1.02, 1.18]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 944 + }, + { + "path": "8297-275155-0013.flac", + "question": "Find the moment 'sit' is clearly audible in the audio.", + "choice_a": "[1.91, 2.26]", + "choice_b": "[2.26, 2.56]", + "choice_c": "[1.19, 1.84]", + "choice_d": "[0.51, 1.00]", + "answer_gt": "[0.51, 1.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 945 + }, + { + "path": "3538-163622-0000.flac", + "question": "Locate the segment where 'asked' is faintly heard.", + "choice_a": "[1.44, 1.70]", + "choice_b": "[1.19, 1.33]", + "choice_c": "[2.49, 2.73]", + "choice_d": "[0.85, 1.19]", + "answer_gt": "[2.49, 2.73]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 946 + }, + { + "path": "4323-13259-0000.flac", + "question": "Establish when 'events' is first mentioned in the clip.", + "choice_a": "[9.27, 9.77]", + "choice_b": "[8.36, 8.72]", + "choice_c": "[1.20, 11.3]", + "choice_d": "[2.03, 2.36]", + "answer_gt": "[8.36, 8.72]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 947 + }, + { + "path": "4852-28312-0010.flac", + "question": "Determine the point in the audio when 'and' is repeated.", + "choice_a": "[1.47, 1.60]", + "choice_b": "[1.69, 1.97]", + "choice_c": "[2.59, 2.90]", + "choice_d": "[0.31, 0.52]", + "answer_gt": "[1.47, 1.60]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 948 + }, + { + "path": "121-127105-0000.flac", + "question": "At what time interval does 'was' occur in the audio clip?", + "choice_a": "[4.54, 4.84]", + "choice_b": "[0.59, 0.83]", + "choice_c": "[4.84, 4.93]", + "choice_d": "[4.93, 5.01]", + "answer_gt": "[0.59, 0.83]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 949 + }, + { + "path": "5142-33396-0016.flac", + "question": "Find the moment 'the' is clearly audible in the audio.", + "choice_a": "[0.94, 1.00]", + "choice_b": "[1.00, 1.34]", + "choice_c": "[0.64, 0.94]", + "choice_d": "[0.31, 0.46]", + "answer_gt": "[0.94, 1.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 950 + }, + { + "path": "6930-81414-0012.flac", + "question": "Listen for the precise interval when 'voice' is dominant.", + "choice_a": "[1.06, 1.20]", + "choice_b": "[1.29, 2.01]", + "choice_c": "[0.68, 1.06]", + "choice_d": "[2.15, 3.01]", + "answer_gt": "[0.68, 1.06]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 951 + }, + { + "path": "6295-244435-0000.flac", + "question": "Determine the point in the audio when 'seven' is repeated.", + "choice_a": "[0.21, 0.49]", + "choice_b": "[0.49, 0.94]", + "choice_c": "[1.72, 1.86]", + "choice_d": "[1.72, 1.86]", + "answer_gt": "[0.49, 0.94]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 952 + }, + { + "path": "3536-8226-0009.flac", + "question": "At what time interval does 'he' occur in the audio clip?", + "choice_a": "[0.23, 0.47]", + "choice_b": "[1.46, 1.68]", + "choice_c": "[1.17, 1.36]", + "choice_d": "[0.75, 1.17]", + "answer_gt": "[0.23, 0.47]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 953 + }, + { + "path": "3570-5694-0016.flac", + "question": "Identify the time segment when 'consumer' is mentioned in the audio.", + "choice_a": "[6.57, 6.84]", + "choice_b": "[8.79, 9.36]", + "choice_c": "[3.60, 3.80]", + "choice_d": "[3.60, 3.80]", + "answer_gt": "[8.79, 9.36]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 954 + }, + { + "path": "6841-88294-0040.flac", + "question": "Find the duration when 'i' occurs consecutively in the audio.", + "choice_a": "[2.09, 2.47]", + "choice_b": "[1.00, 1.08]", + "choice_c": "[1.85, 1.94]", + "choice_d": "[1.00, 1.08]", + "answer_gt": "[1.00, 1.08]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 955 + }, + { + "path": "3764-168671-0007.flac", + "question": "Find out when the speaker refers to 'hinted' during the audio.", + "choice_a": "[2.97, 3.08]", + "choice_b": "[6.36, 6.97]", + "choice_c": "[4.42, 4.71]", + "choice_d": "[4.71, 4.87]", + "answer_gt": "[4.42, 4.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 956 + }, + { + "path": "700-122868-0008.flac", + "question": "At what interval does 'they' appear in the audio?", + "choice_a": "[1.06, 1.28]", + "choice_b": "[0.53, 0.65]", + "choice_c": "[0.65, 0.73]", + "choice_d": "[2.73, 2.85]", + "answer_gt": "[0.53, 0.65]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 957 + }, + { + "path": "2830-3980-0018.flac", + "question": "At what interval does 'of' appear in the audio?", + "choice_a": "[0.50, 0.67]", + "choice_b": "[1.20, 1.70]", + "choice_c": "[2.29, 2.39]", + "choice_d": "[1.70, 1.78]", + "answer_gt": "[2.29, 2.39]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 958 + }, + { + "path": "7641-96252-0017.flac", + "question": "Find out when the speaker refers to 'giordano' during the audio.", + "choice_a": "[4.78, 4.92]", + "choice_b": "[1.95, 2.74]", + "choice_c": "[4.39, 4.78]", + "choice_d": "[5.73, 6.00]", + "answer_gt": "[1.95, 2.74]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 959 + }, + { + "path": "3536-8226-0027.flac", + "question": "Discover when 'all' is introduced in the conversation.", + "choice_a": "[0.62, 0.86]", + "choice_b": "[0.42, 0.57]", + "choice_c": "[1.17, 1.29]", + "choice_d": "[0.86, 1.17]", + "answer_gt": "[0.62, 0.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 960 + }, + { + "path": "2412-153954-0019.flac", + "question": "Ascertain the exact time 'which' is mentioned in the audio.", + "choice_a": "[2.35, 2.51]", + "choice_b": "[3.26, 3.40]", + "choice_c": "[2.51, 2.63]", + "choice_d": "[5.27, 5.84]", + "answer_gt": "[2.35, 2.51]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 961 + }, + { + "path": "5105-28241-0011.flac", + "question": "Determine the point in the audio when 'only' is repeated.", + "choice_a": "[6.76, 7.02]", + "choice_b": "[5.93, 6.47]", + "choice_c": "[0.43, 0.67]", + "choice_d": "[0.67, 0.81]", + "answer_gt": "[6.76, 7.02]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 962 + }, + { + "path": "1650-157641-0010.flac", + "question": "Listen for the occurrence of 'lay' in the recording.", + "choice_a": "[0.88, 1.20]", + "choice_b": "[1.89, 2.25]", + "choice_c": "[5.04, 5.11]", + "choice_d": "[1.71, 1.89]", + "answer_gt": "[0.88, 1.20]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 963 + }, + { + "path": "3005-163399-0021.flac", + "question": "When is 'dreadful' first mentioned in the clip?", + "choice_a": "[3.96, 4.52]", + "choice_b": "[1.60, 1.81]", + "choice_c": "[0.70, 1.18]", + "choice_d": "[1.50, 1.60]", + "answer_gt": "[3.96, 4.52]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 964 + }, + { + "path": "5536-43359-0004.flac", + "question": "Find out the time when 'due' is immediately occured.", + "choice_a": "[1.84, 1.93]", + "choice_b": "[1.13, 1.19]", + "choice_c": "[3.63, 3.98]", + "choice_d": "[0.57, 0.71]", + "answer_gt": "[0.57, 0.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 965 + }, + { + "path": "6829-68769-0043.flac", + "question": "Recognize the time when 'deserves' is mentioned by the speaker.", + "choice_a": "[0.15, 0.28]", + "choice_b": "[0.84, 0.87]", + "choice_c": "[1.31, 1.68]", + "choice_d": "[0.37, 0.84]", + "answer_gt": "[0.37, 0.84]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 966 + }, + { + "path": "8555-292519-0013.flac", + "question": "Locate the segment where 'in' is faintly heard.", + "choice_a": "[1.87, 2.02]", + "choice_b": "[2.95, 3.04]", + "choice_c": "[0.49, 0.75]", + "choice_d": "[1.22, 1.84]", + "answer_gt": "[2.95, 3.04]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 967 + }, + { + "path": "5442-41169-0010.flac", + "question": "Find the duration when 'up' occurs consecutively in the audio.", + "choice_a": "[1.97, 2.14]", + "choice_b": "[1.77, 1.97]", + "choice_c": "[1.55, 1.77]", + "choice_d": "[0.19, 0.54]", + "answer_gt": "[1.97, 2.14]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 968 + }, + { + "path": "7641-96670-0000.flac", + "question": "Recognize the time when 'said' is most clearly articulated.", + "choice_a": "[3.61, 3.81]", + "choice_b": "[0.19, 0.35]", + "choice_c": "[1.85, 2.26]", + "choice_d": "[2.54, 2.90]", + "answer_gt": "[1.85, 2.26]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 969 + }, + { + "path": "1272-135031-0017.flac", + "question": "Determine the point in the audio when 'sharp' is discussed.", + "choice_a": "[4.96, 5.11]", + "choice_b": "[2.21, 2.71]", + "choice_c": "[1.94, 2.00]", + "choice_d": "[4.07, 4.19]", + "answer_gt": "[2.21, 2.71]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 970 + }, + { + "path": "3997-180297-0025.flac", + "question": "Locate the part of the audio where 'her' is expressed.", + "choice_a": "[2.16, 2.62]", + "choice_b": "[4.20, 4.40]", + "choice_c": "[0.53, 1.10]", + "choice_d": "[2.09, 2.16]", + "answer_gt": "[2.09, 2.16]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 971 + }, + { + "path": "367-293981-0005.flac", + "question": "Determine the point in the audio when 'line' is discussed.", + "choice_a": "[2.19, 2.54]", + "choice_b": "[0.44, 0.57]", + "choice_c": "[0.57, 0.83]", + "choice_d": "[0.19, 0.44]", + "answer_gt": "[2.19, 2.54]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 972 + }, + { + "path": "3000-15664-0043.flac", + "question": "Listen for when 'shasta' is emphasized by the narrator.", + "choice_a": "[8.56, 8.93]", + "choice_b": "[1.53, 2.21]", + "choice_c": "[4.73, 4.84]", + "choice_d": "[1.19, 1.53]", + "answer_gt": "[1.53, 2.21]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 973 + }, + { + "path": "8254-115543-0026.flac", + "question": "Recognize the time when 'it' is most clearly articulated.", + "choice_a": "[0.77, 0.86]", + "choice_b": "[1.82, 1.90]", + "choice_c": "[0.86, 1.22]", + "choice_d": "[1.53, 1.69]", + "answer_gt": "[0.77, 0.86]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 974 + }, + { + "path": "116-288048-0025.flac", + "question": "Find out the time when 'no' is immediately occured.", + "choice_a": "[9.88, 10.5]", + "choice_b": "[4.13, 4.33]", + "choice_c": "[4.33, 4.80]", + "choice_d": "[4.13, 4.33]", + "answer_gt": "[4.13, 4.33]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 975 + }, + { + "path": "5543-27761-0023.flac", + "question": "Determine the point in the audio when 'nearer' is discussed.", + "choice_a": "[1.27, 1.63]", + "choice_b": "[2.11, 2.48]", + "choice_c": "[0.54, 1.13]", + "choice_d": "[1.63, 2.00]", + "answer_gt": "[1.63, 2.00]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 976 + }, + { + "path": "5849-50962-0001.flac", + "question": "Pinpoint the exact moment 'to' is repeated in the audio.", + "choice_a": "[2.15, 2.28]", + "choice_b": "[4.05, 4.33]", + "choice_c": "[0.56, 0.61]", + "choice_d": "[4.33, 4.58]", + "answer_gt": "[2.15, 2.28]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 977 + }, + { + "path": "6313-66125-0014.flac", + "question": "Note the time interval that includes the mention of 'swaying'.", + "choice_a": "[1.12, 1.28]", + "choice_b": "[0.26, 0.34]", + "choice_c": "[1.65, 2.18]", + "choice_d": "[2.53, 2.69]", + "answer_gt": "[1.65, 2.18]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 978 + }, + { + "path": "8280-266249-0003.flac", + "question": "Identify the moment 'and' is mentioned in the audio.", + "choice_a": "[5.62, 5.80]", + "choice_b": "[7.69, 8.15]", + "choice_c": "[0.82, 11.2]", + "choice_d": "[7.36, 7.69]", + "answer_gt": "[5.62, 5.80]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 979 + }, + { + "path": "3729-6852-0040.flac", + "question": "Listen for the precise interval when 'you' is dominant.", + "choice_a": "[4.08, 4.25]", + "choice_b": "[3.22, 3.63]", + "choice_c": "[4.25, 4.47]", + "choice_d": "[4.08, 4.25]", + "answer_gt": "[4.08, 4.25]", + "task_name": "Speech_Grounding", + "dataset_name": "librispeech", + "uniq_id": 980 + }, + { + "path": "common_voice_fr_19680269.mp3", + "question": "identify the language of this speech.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 981 + }, + { + "path": "common_voice_en_19960977.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 982 + }, + { + "path": "common_voice_de_18792412.mp3", + "question": "What language is used in this speech?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 983 + }, + { + "path": "common_voice_en_19663617.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 984 + }, + { + "path": "common_voice_en_19106358.mp3", + "question": "Identify the language of the speech.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 985 + }, + { + "path": "common_voice_de_19727936.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 986 + }, + { + "path": "common_voice_en_403936.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 987 + }, + { + "path": "common_voice_en_545581.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 988 + }, + { + "path": "common_voice_en_19952829.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 989 + }, + { + "path": "common_voice_en_19973541.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 990 + }, + { + "path": "common_voice_de_19167467.mp3", + "question": "What language is the speaker using?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 991 + }, + { + "path": "common_voice_fr_19600742.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 992 + }, + { + "path": "common_voice_en_32845.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 993 + }, + { + "path": "common_voice_es_19639735.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 994 + }, + { + "path": "common_voice_en_19825847.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 995 + }, + { + "path": "common_voice_fr_19614583.mp3", + "question": "Which language is used in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 996 + }, + { + "path": "common_voice_fr_18325103.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 997 + }, + { + "path": "common_voice_es_19949530.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 998 + }, + { + "path": "common_voice_en_19611318.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 999 + }, + { + "path": "common_voice_en_195779.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1000 + }, + { + "path": "common_voice_en_19628916.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1001 + }, + { + "path": "common_voice_fr_19754444.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1002 + }, + { + "path": "common_voice_en_19693022.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1003 + }, + { + "path": "common_voice_de_18481673.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1004 + }, + { + "path": "common_voice_zh-CN_18789820.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1005 + }, + { + "path": "common_voice_de_19410018.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1006 + }, + { + "path": "common_voice_fr_18043859.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1007 + }, + { + "path": "common_voice_es_19695490.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1008 + }, + { + "path": "common_voice_fr_18302402.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1009 + }, + { + "path": "common_voice_en_15447980.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1010 + }, + { + "path": "common_voice_en_507964.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1011 + }, + { + "path": "common_voice_en_656407.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1012 + }, + { + "path": "common_voice_fr_18619543.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1013 + }, + { + "path": "common_voice_en_19649069.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1014 + }, + { + "path": "common_voice_it_19972607.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1015 + }, + { + "path": "common_voice_fr_18600240.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1016 + }, + { + "path": "common_voice_it_19990185.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1017 + }, + { + "path": "common_voice_it_18981152.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1018 + }, + { + "path": "common_voice_de_18018666.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1019 + }, + { + "path": "common_voice_en_19627678.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1020 + }, + { + "path": "common_voice_en_18883695.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1021 + }, + { + "path": "common_voice_en_19867946.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1022 + }, + { + "path": "common_voice_it_20045766.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1023 + }, + { + "path": "common_voice_en_208709.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1024 + }, + { + "path": "common_voice_en_18553506.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1025 + }, + { + "path": "common_voice_en_19723815.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1026 + }, + { + "path": "common_voice_en_18484513.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1027 + }, + { + "path": "common_voice_en_19718936.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1028 + }, + { + "path": "common_voice_de_17331770.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1029 + }, + { + "path": "common_voice_de_19577457.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1030 + }, + { + "path": "common_voice_it_17859889.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1031 + }, + { + "path": "common_voice_zh-CN_18682360.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1032 + }, + { + "path": "common_voice_en_19754513.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1033 + }, + { + "path": "common_voice_de_18309447.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1034 + }, + { + "path": "common_voice_de_18881695.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1035 + }, + { + "path": "common_voice_fr_19752866.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1036 + }, + { + "path": "common_voice_en_686468.mp3", + "question": "In which language was the announcement made?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1037 + }, + { + "path": "common_voice_en_20143281.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1038 + }, + { + "path": "common_voice_es_19608994.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1039 + }, + { + "path": "common_voice_de_17848918.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1040 + }, + { + "path": "common_voice_es_20039218.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1041 + }, + { + "path": "common_voice_en_19528390.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1042 + }, + { + "path": "common_voice_es_19730325.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1043 + }, + { + "path": "common_voice_en_19793465.mp3", + "question": "What language is used in this speech?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1044 + }, + { + "path": "common_voice_de_18227397.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1045 + }, + { + "path": "common_voice_fr_19733920.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1046 + }, + { + "path": "common_voice_de_17587538.mp3", + "question": "Guess the language in the audio.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1047 + }, + { + "path": "common_voice_it_19912576.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1048 + }, + { + "path": "common_voice_de_19173110.mp3", + "question": "In which language was the announcement made?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1049 + }, + { + "path": "common_voice_es_19601414.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1050 + }, + { + "path": "common_voice_de_20003115.mp3", + "question": "In which language is this speech being given?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1051 + }, + { + "path": "common_voice_de_18461277.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1052 + }, + { + "path": "common_voice_es_19627992.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1053 + }, + { + "path": "common_voice_de_17527243.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1054 + }, + { + "path": "common_voice_zh-CN_19005255.mp3", + "question": "What language is the speech in?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1055 + }, + { + "path": "common_voice_en_559610.mp3", + "question": "The speaker is using which language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1056 + }, + { + "path": "common_voice_zh-CN_18637715.mp3", + "question": "The speech was in which language?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1057 + }, + { + "path": "common_voice_en_19844537.mp3", + "question": "In which language was the announcement made?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1058 + }, + { + "path": "common_voice_es_19666728.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1059 + }, + { + "path": "common_voice_en_19757018.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1060 + }, + { + "path": "common_voice_en_19955413.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1061 + }, + { + "path": "common_voice_fr_19421328.mp3", + "question": "The speech was in which language?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1062 + }, + { + "path": "common_voice_it_20036440.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1063 + }, + { + "path": "common_voice_fr_19736741.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1064 + }, + { + "path": "common_voice_en_20056091.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1065 + }, + { + "path": "common_voice_en_478162.mp3", + "question": "The speaker is using which language?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1066 + }, + { + "path": "common_voice_en_19814547.mp3", + "question": "In which language was the announcement made?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1067 + }, + { + "path": "common_voice_de_18093657.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1068 + }, + { + "path": "common_voice_de_18637308.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1069 + }, + { + "path": "common_voice_fr_19649973.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1070 + }, + { + "path": "common_voice_fr_19757377.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1071 + }, + { + "path": "common_voice_en_57493.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1072 + }, + { + "path": "common_voice_de_18227917.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1073 + }, + { + "path": "common_voice_fr_19686845.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1074 + }, + { + "path": "common_voice_de_19673876.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1075 + }, + { + "path": "common_voice_it_19972265.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1076 + }, + { + "path": "common_voice_de_18543256.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1077 + }, + { + "path": "common_voice_es_19943635.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1078 + }, + { + "path": "common_voice_en_17902010.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1079 + }, + { + "path": "common_voice_fr_17359412.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1080 + }, + { + "path": "common_voice_en_19950629.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1081 + }, + { + "path": "common_voice_en_554042.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1082 + }, + { + "path": "common_voice_en_19999352.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1083 + }, + { + "path": "common_voice_en_17250737.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1084 + }, + { + "path": "common_voice_en_20021087.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1085 + }, + { + "path": "common_voice_en_19608603.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1086 + }, + { + "path": "common_voice_en_19490061.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1087 + }, + { + "path": "common_voice_en_19730255.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1088 + }, + { + "path": "common_voice_fr_19619707.mp3", + "question": "The speech was presented in what language?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1089 + }, + { + "path": "common_voice_es_19966482.mp3", + "question": "In which language is this speech being given?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1090 + }, + { + "path": "common_voice_fr_19798524.mp3", + "question": "What language is the speaker using?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1091 + }, + { + "path": "common_voice_en_18487929.mp3", + "question": "In which language was the announcement made?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1092 + }, + { + "path": "common_voice_en_19955987.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1093 + }, + { + "path": "common_voice_it_19753185.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1094 + }, + { + "path": "common_voice_fr_17787697.mp3", + "question": "In which language was the announcement made?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1095 + }, + { + "path": "common_voice_en_19675069.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1096 + }, + { + "path": "common_voice_it_20036112.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1097 + }, + { + "path": "common_voice_fr_19131914.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1098 + }, + { + "path": "common_voice_de_19534359.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1099 + }, + { + "path": "common_voice_it_19989445.mp3", + "question": "The speaker is using which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1100 + }, + { + "path": "common_voice_en_19955986.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1101 + }, + { + "path": "common_voice_fr_19867579.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1102 + }, + { + "path": "common_voice_en_20050615.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1103 + }, + { + "path": "common_voice_de_20014841.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1104 + }, + { + "path": "common_voice_es_19667832.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1105 + }, + { + "path": "common_voice_en_17274621.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1106 + }, + { + "path": "common_voice_de_18521595.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1107 + }, + { + "path": "common_voice_en_19652974.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1108 + }, + { + "path": "common_voice_it_19989475.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1109 + }, + { + "path": "common_voice_en_19016882.mp3", + "question": "The speech was in which language?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1110 + }, + { + "path": "common_voice_de_18775068.mp3", + "question": "The speech was in which language?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1111 + }, + { + "path": "common_voice_en_19101813.mp3", + "question": "Which language is used in this speech?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1112 + }, + { + "path": "common_voice_en_19718214.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1113 + }, + { + "path": "common_voice_en_18935703.mp3", + "question": "In which language is this speech being given?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1114 + }, + { + "path": "common_voice_es_19972013.mp3", + "question": "Identify the language from the speech.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1115 + }, + { + "path": "common_voice_de_19571302.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1116 + }, + { + "path": "common_voice_en_19667039.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1117 + }, + { + "path": "common_voice_en_17815403.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1118 + }, + { + "path": "common_voice_en_426620.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1119 + }, + { + "path": "common_voice_en_125163.mp3", + "question": "What language is the speech in?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1120 + }, + { + "path": "common_voice_fr_19680707.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1121 + }, + { + "path": "common_voice_en_19673504.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1122 + }, + { + "path": "common_voice_en_18669616.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1123 + }, + { + "path": "common_voice_de_19741434.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1124 + }, + { + "path": "common_voice_en_19964953.mp3", + "question": "Identify the language from the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1125 + }, + { + "path": "common_voice_de_19847031.mp3", + "question": "Guess the language in the audio.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1126 + }, + { + "path": "common_voice_de_19354362.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1127 + }, + { + "path": "common_voice_fr_19824301.mp3", + "question": "Identify the language of the speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1128 + }, + { + "path": "common_voice_de_18236266.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1129 + }, + { + "path": "common_voice_es_18355459.mp3", + "question": "The speech was presented in what language?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1130 + }, + { + "path": "common_voice_fr_17756109.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1131 + }, + { + "path": "common_voice_de_17339402.mp3", + "question": "What language is the speech in?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1132 + }, + { + "path": "common_voice_es_19767492.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1133 + }, + { + "path": "common_voice_en_18257271.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1134 + }, + { + "path": "common_voice_es_19714780.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1135 + }, + { + "path": "common_voice_fr_19599498.mp3", + "question": "What language is used in this speech?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1136 + }, + { + "path": "common_voice_en_20062010.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1137 + }, + { + "path": "common_voice_es_19639113.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1138 + }, + { + "path": "common_voice_en_19660281.mp3", + "question": "What language is the speech in?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1139 + }, + { + "path": "common_voice_en_18125376.mp3", + "question": "Guess the language of the speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1140 + }, + { + "path": "common_voice_zh-CN_18983162.mp3", + "question": "What language is the speech in?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1141 + }, + { + "path": "common_voice_en_19754389.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1142 + }, + { + "path": "common_voice_en_18533903.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1143 + }, + { + "path": "common_voice_en_482634.mp3", + "question": "Which language is used in this speech?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1144 + }, + { + "path": "common_voice_en_19948420.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1145 + }, + { + "path": "common_voice_es_19699654.mp3", + "question": "identify the language of this speech.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1146 + }, + { + "path": "common_voice_en_20027352.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1147 + }, + { + "path": "common_voice_en_18340361.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1148 + }, + { + "path": "common_voice_de_18928297.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1149 + }, + { + "path": "common_voice_en_210167.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1150 + }, + { + "path": "common_voice_de_19675395.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1151 + }, + { + "path": "common_voice_es_19954268.mp3", + "question": "Guess the language of the speech.", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1152 + }, + { + "path": "common_voice_it_19993252.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1153 + }, + { + "path": "common_voice_it_19985601.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1154 + }, + { + "path": "common_voice_en_1416459.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1155 + }, + { + "path": "common_voice_fr_18170324.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1156 + }, + { + "path": "common_voice_en_19715498.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1157 + }, + { + "path": "common_voice_es_19611831.mp3", + "question": "Identify the language of the speech.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1158 + }, + { + "path": "common_voice_de_17519104.mp3", + "question": "The speaker is using which language?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1159 + }, + { + "path": "common_voice_en_19718213.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1160 + }, + { + "path": "common_voice_en_20002401.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1161 + }, + { + "path": "common_voice_de_19954711.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1162 + }, + { + "path": "common_voice_de_18205935.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1163 + }, + { + "path": "common_voice_en_17940817.mp3", + "question": "identify the language of this speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1164 + }, + { + "path": "common_voice_en_18732130.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1165 + }, + { + "path": "common_voice_en_19805694.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1166 + }, + { + "path": "common_voice_en_20039940.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1167 + }, + { + "path": "common_voice_en_19607147.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1168 + }, + { + "path": "common_voice_es_19740928.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1169 + }, + { + "path": "common_voice_fr_18073109.mp3", + "question": "What language is the speaker using?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1170 + }, + { + "path": "common_voice_en_17879878.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1171 + }, + { + "path": "common_voice_es_19942409.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1172 + }, + { + "path": "common_voice_es_19632626.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1173 + }, + { + "path": "common_voice_en_19684357.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1174 + }, + { + "path": "common_voice_en_20088061.mp3", + "question": "Guess the language in the audio.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1175 + }, + { + "path": "common_voice_es_19681908.mp3", + "question": "What language is the speaker using?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1176 + }, + { + "path": "common_voice_en_88831.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1177 + }, + { + "path": "common_voice_fr_19702783.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1178 + }, + { + "path": "common_voice_en_19957490.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1179 + }, + { + "path": "common_voice_en_20038776.mp3", + "question": "What language is the speaker using?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1180 + }, + { + "path": "common_voice_fr_19595100.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1181 + }, + { + "path": "common_voice_es_19660130.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1182 + }, + { + "path": "common_voice_de_17836056.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1183 + }, + { + "path": "common_voice_de_19675362.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1184 + }, + { + "path": "common_voice_en_19652898.mp3", + "question": "In which language is this speech being given?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1185 + }, + { + "path": "common_voice_en_18973476.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1186 + }, + { + "path": "common_voice_en_516323.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1187 + }, + { + "path": "common_voice_de_17652905.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1188 + }, + { + "path": "common_voice_fr_18150063.mp3", + "question": "The speech was presented in what language?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1189 + }, + { + "path": "common_voice_fr_18235800.mp3", + "question": "Identify the language from the speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1190 + }, + { + "path": "common_voice_en_19755754.mp3", + "question": "In which language was the announcement made?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1191 + }, + { + "path": "common_voice_en_69226.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1192 + }, + { + "path": "common_voice_de_19167469.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1193 + }, + { + "path": "common_voice_it_20057297.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1194 + }, + { + "path": "common_voice_en_19878657.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1195 + }, + { + "path": "common_voice_es_19625951.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1196 + }, + { + "path": "common_voice_de_19728017.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1197 + }, + { + "path": "common_voice_en_536652.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1198 + }, + { + "path": "common_voice_en_17275508.mp3", + "question": "Guess the language in the audio.", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1199 + }, + { + "path": "common_voice_en_141797.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1200 + }, + { + "path": "common_voice_de_17497461.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1201 + }, + { + "path": "common_voice_it_20042688.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1202 + }, + { + "path": "common_voice_en_19723484.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1203 + }, + { + "path": "common_voice_es_18705867.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1204 + }, + { + "path": "common_voice_es_19691179.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1205 + }, + { + "path": "common_voice_en_19399442.mp3", + "question": "In which language was the announcement made?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1206 + }, + { + "path": "common_voice_en_19801837.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1207 + }, + { + "path": "common_voice_en_586122.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1208 + }, + { + "path": "common_voice_en_187462.mp3", + "question": "The speaker is using which language?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1209 + }, + { + "path": "common_voice_de_19150706.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1210 + }, + { + "path": "common_voice_en_19770420.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1211 + }, + { + "path": "common_voice_en_19713333.mp3", + "question": "The speaker is using which language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1212 + }, + { + "path": "common_voice_it_20005067.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1213 + }, + { + "path": "common_voice_en_2542578.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1214 + }, + { + "path": "common_voice_es_19696105.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1215 + }, + { + "path": "common_voice_en_17699273.mp3", + "question": "The speech was presented in what language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1216 + }, + { + "path": "common_voice_fr_18372457.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1217 + }, + { + "path": "common_voice_en_615156.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1218 + }, + { + "path": "common_voice_it_19970833.mp3", + "question": "In which language was the announcement made?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1219 + }, + { + "path": "common_voice_en_18638082.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1220 + }, + { + "path": "common_voice_zh-CN_18768110.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1221 + }, + { + "path": "common_voice_en_149358.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1222 + }, + { + "path": "common_voice_es_19649904.mp3", + "question": "The speech was presented in what language?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1223 + }, + { + "path": "common_voice_de_19130846.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1224 + }, + { + "path": "common_voice_en_36534.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1225 + }, + { + "path": "common_voice_es_19613360.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1226 + }, + { + "path": "common_voice_es_19698891.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1227 + }, + { + "path": "common_voice_en_8774093.mp3", + "question": "identify the language of this speech.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1228 + }, + { + "path": "common_voice_de_19154350.mp3", + "question": "What language is the speaker using?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1229 + }, + { + "path": "common_voice_fr_17730768.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1230 + }, + { + "path": "common_voice_en_137103.mp3", + "question": "The speech was presented in what language?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1231 + }, + { + "path": "common_voice_zh-CN_18995033.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1232 + }, + { + "path": "common_voice_es_20063316.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1233 + }, + { + "path": "common_voice_en_20050453.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1234 + }, + { + "path": "common_voice_en_34067.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1235 + }, + { + "path": "common_voice_it_18783386.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1236 + }, + { + "path": "common_voice_it_19999854.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1237 + }, + { + "path": "common_voice_es_19652151.mp3", + "question": "In which language was the announcement made?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1238 + }, + { + "path": "common_voice_en_19980731.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1239 + }, + { + "path": "common_voice_en_417460.mp3", + "question": "What language is the speaker using?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1240 + }, + { + "path": "common_voice_en_590648.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1241 + }, + { + "path": "common_voice_en_667214.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1242 + }, + { + "path": "common_voice_en_19763272.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1243 + }, + { + "path": "common_voice_de_18476364.mp3", + "question": "Identify the language from the speech.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1244 + }, + { + "path": "common_voice_en_17873663.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1245 + }, + { + "path": "common_voice_es_19723778.mp3", + "question": "identify the language of this speech.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1246 + }, + { + "path": "common_voice_de_18188172.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1247 + }, + { + "path": "common_voice_en_19601585.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1248 + }, + { + "path": "common_voice_es_19958374.mp3", + "question": "The speaker is using which language?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1249 + }, + { + "path": "common_voice_en_212507.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1250 + }, + { + "path": "common_voice_it_17704567.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1251 + }, + { + "path": "common_voice_de_18210798.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1252 + }, + { + "path": "common_voice_en_19714376.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1253 + }, + { + "path": "common_voice_it_19996602.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1254 + }, + { + "path": "common_voice_en_18687447.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1255 + }, + { + "path": "common_voice_fr_19643208.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1256 + }, + { + "path": "common_voice_de_18494522.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1257 + }, + { + "path": "common_voice_en_18306107.mp3", + "question": "The speech was presented in what language?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1258 + }, + { + "path": "common_voice_fr_18395756.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1259 + }, + { + "path": "common_voice_de_18204584.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1260 + }, + { + "path": "common_voice_fr_19080420.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1261 + }, + { + "path": "common_voice_de_19464290.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1262 + }, + { + "path": "common_voice_it_19971213.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1263 + }, + { + "path": "common_voice_es_19606324.mp3", + "question": "The speech was in which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1264 + }, + { + "path": "common_voice_es_19722587.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1265 + }, + { + "path": "common_voice_en_19695573.mp3", + "question": "Guess the language in the audio.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1266 + }, + { + "path": "common_voice_fr_19669887.mp3", + "question": "The speech was in which language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1267 + }, + { + "path": "common_voice_en_19747527.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1268 + }, + { + "path": "common_voice_en_19748159.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1269 + }, + { + "path": "common_voice_es_18666622.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1270 + }, + { + "path": "common_voice_it_20054743.mp3", + "question": "The speech was in which language?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1271 + }, + { + "path": "common_voice_fr_17309878.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1272 + }, + { + "path": "common_voice_en_19998298.mp3", + "question": "The speaker is using which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1273 + }, + { + "path": "common_voice_zh-CN_19483502.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1274 + }, + { + "path": "common_voice_de_17741712.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1275 + }, + { + "path": "common_voice_fr_18004519.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1276 + }, + { + "path": "common_voice_fr_18935926.mp3", + "question": "Identify the language from the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1277 + }, + { + "path": "common_voice_de_19182466.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1278 + }, + { + "path": "common_voice_it_18137730.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1279 + }, + { + "path": "common_voice_es_19669291.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1280 + }, + { + "path": "common_voice_en_18401387.mp3", + "question": "In which language is this speech being given?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1281 + }, + { + "path": "common_voice_en_19965461.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1282 + }, + { + "path": "common_voice_de_18726589.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1283 + }, + { + "path": "common_voice_fr_19967683.mp3", + "question": "The speech was presented in what language?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1284 + }, + { + "path": "common_voice_es_19954817.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1285 + }, + { + "path": "common_voice_fr_19176902.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1286 + }, + { + "path": "common_voice_de_19158064.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1287 + }, + { + "path": "common_voice_fr_19701849.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1288 + }, + { + "path": "common_voice_ja_19485633.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "ja", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1289 + }, + { + "path": "common_voice_zh-CN_18689594.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1290 + }, + { + "path": "common_voice_en_18492670.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1291 + }, + { + "path": "common_voice_de_18206560.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1292 + }, + { + "path": "common_voice_en_18508595.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1293 + }, + { + "path": "common_voice_en_18754788.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1294 + }, + { + "path": "common_voice_de_19196611.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1295 + }, + { + "path": "common_voice_es_19611281.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1296 + }, + { + "path": "common_voice_fr_18196613.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1297 + }, + { + "path": "common_voice_en_474108.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1298 + }, + { + "path": "common_voice_de_19552034.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1299 + }, + { + "path": "common_voice_en_20062959.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1300 + }, + { + "path": "common_voice_en_606849.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1301 + }, + { + "path": "common_voice_fr_19274887.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1302 + }, + { + "path": "common_voice_en_508133.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1303 + }, + { + "path": "common_voice_en_79491.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1304 + }, + { + "path": "common_voice_en_101984.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1305 + }, + { + "path": "common_voice_de_19363341.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1306 + }, + { + "path": "common_voice_es_19634649.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1307 + }, + { + "path": "common_voice_es_19962814.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1308 + }, + { + "path": "common_voice_en_18466092.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1309 + }, + { + "path": "common_voice_fr_20041637.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1310 + }, + { + "path": "common_voice_es_19696673.mp3", + "question": "What language is the speech in?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1311 + }, + { + "path": "common_voice_it_19469105.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1312 + }, + { + "path": "common_voice_it_17981493.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1313 + }, + { + "path": "common_voice_en_19717360.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1314 + }, + { + "path": "common_voice_it_20013631.mp3", + "question": "The speaker is using which language?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1315 + }, + { + "path": "common_voice_en_19573594.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1316 + }, + { + "path": "common_voice_fr_19642811.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1317 + }, + { + "path": "common_voice_en_221271.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1318 + }, + { + "path": "common_voice_de_19548609.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1319 + }, + { + "path": "common_voice_ja_19682962.mp3", + "question": "What language is used in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "ja", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1320 + }, + { + "path": "common_voice_fr_19979244.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1321 + }, + { + "path": "common_voice_en_293710.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1322 + }, + { + "path": "common_voice_fr_19602790.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1323 + }, + { + "path": "common_voice_fr_18094417.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1324 + }, + { + "path": "common_voice_fr_20003525.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1325 + }, + { + "path": "common_voice_it_20004489.mp3", + "question": "Identify the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1326 + }, + { + "path": "common_voice_en_103457.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1327 + }, + { + "path": "common_voice_fr_18018259.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1328 + }, + { + "path": "common_voice_en_17265230.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1329 + }, + { + "path": "common_voice_de_19131984.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1330 + }, + { + "path": "common_voice_es_19667128.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1331 + }, + { + "path": "common_voice_zh-CN_18600904.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1332 + }, + { + "path": "common_voice_fr_19632834.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1333 + }, + { + "path": "common_voice_en_19382382.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1334 + }, + { + "path": "common_voice_de_18219365.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1335 + }, + { + "path": "common_voice_zh-CN_20036214.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1336 + }, + { + "path": "common_voice_de_20010420.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1337 + }, + { + "path": "common_voice_en_5877336.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "de", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1338 + }, + { + "path": "common_voice_de_19856421.mp3", + "question": "The speaker is using which language?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1339 + }, + { + "path": "common_voice_es_19607350.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1340 + }, + { + "path": "common_voice_es_19584617.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1341 + }, + { + "path": "common_voice_en_19848348.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1342 + }, + { + "path": "common_voice_fr_19661299.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1343 + }, + { + "path": "common_voice_en_19669545.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1344 + }, + { + "path": "common_voice_en_508132.mp3", + "question": "In which language is this speech being given?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1345 + }, + { + "path": "common_voice_en_19670349.mp3", + "question": "Which language is used in this speech?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1346 + }, + { + "path": "common_voice_en_19982852.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1347 + }, + { + "path": "common_voice_es_19634014.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1348 + }, + { + "path": "common_voice_it_19995798.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1349 + }, + { + "path": "common_voice_en_18878076.mp3", + "question": "What language is the speaker using?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1350 + }, + { + "path": "common_voice_en_4776612.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1351 + }, + { + "path": "common_voice_de_19731120.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1352 + }, + { + "path": "common_voice_es_19610795.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1353 + }, + { + "path": "common_voice_en_19964804.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1354 + }, + { + "path": "common_voice_en_19715805.mp3", + "question": "What language is the speaker using?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1355 + }, + { + "path": "common_voice_it_19983812.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1356 + }, + { + "path": "common_voice_fr_19631438.mp3", + "question": "What language is the speech in?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1357 + }, + { + "path": "common_voice_en_19962792.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1358 + }, + { + "path": "common_voice_fr_18954994.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1359 + }, + { + "path": "common_voice_es_19609543.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1360 + }, + { + "path": "common_voice_zh-CN_18979025.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1361 + }, + { + "path": "common_voice_en_667490.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1362 + }, + { + "path": "common_voice_fr_20043356.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1363 + }, + { + "path": "common_voice_fr_18372113.mp3", + "question": "What language is the speaker using?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1364 + }, + { + "path": "common_voice_en_18260944.mp3", + "question": "Guess the language in the audio.", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1365 + }, + { + "path": "common_voice_it_19969700.mp3", + "question": "In which language was the announcement made?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1366 + }, + { + "path": "common_voice_en_17296485.mp3", + "question": "The speech was in which language?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1367 + }, + { + "path": "common_voice_en_19947950.mp3", + "question": "Identify the language from the speech.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1368 + }, + { + "path": "common_voice_en_19150079.mp3", + "question": "In which language is this speech being given?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1369 + }, + { + "path": "common_voice_es_19740482.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1370 + }, + { + "path": "common_voice_fr_19971527.mp3", + "question": "The speaker is using which language?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1371 + }, + { + "path": "common_voice_fr_19962476.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1372 + }, + { + "path": "common_voice_en_10410522.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1373 + }, + { + "path": "common_voice_de_19602213.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1374 + }, + { + "path": "common_voice_es_19669581.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1375 + }, + { + "path": "common_voice_en_18985254.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1376 + }, + { + "path": "common_voice_en_610649.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1377 + }, + { + "path": "common_voice_it_18530057.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1378 + }, + { + "path": "common_voice_de_19164269.mp3", + "question": "identify the language of this speech.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1379 + }, + { + "path": "common_voice_en_20724.mp3", + "question": "What language is the speaker using?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1380 + }, + { + "path": "common_voice_en_19765328.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1381 + }, + { + "path": "common_voice_es_19602979.mp3", + "question": "The speech was in which language?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1382 + }, + { + "path": "common_voice_es_19632335.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1383 + }, + { + "path": "common_voice_en_19718100.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1384 + }, + { + "path": "common_voice_es_19723412.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1385 + }, + { + "path": "common_voice_en_635233.mp3", + "question": "The speaker is using which language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1386 + }, + { + "path": "common_voice_zh-CN_19416775.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1387 + }, + { + "path": "common_voice_es_19717015.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1388 + }, + { + "path": "common_voice_fr_18150059.mp3", + "question": "The speaker is using which language?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1389 + }, + { + "path": "common_voice_fr_19769064.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1390 + }, + { + "path": "common_voice_en_20067344.mp3", + "question": "Identify the language of the speech.", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1391 + }, + { + "path": "common_voice_en_18654359.mp3", + "question": "In which language is this speech being given?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1392 + }, + { + "path": "common_voice_zh-CN_18592607.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1393 + }, + { + "path": "common_voice_en_19164062.mp3", + "question": "Identify the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1394 + }, + { + "path": "common_voice_en_19665694.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1395 + }, + { + "path": "common_voice_it_20020105.mp3", + "question": "The speech was in which language?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1396 + }, + { + "path": "common_voice_en_544962.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1397 + }, + { + "path": "common_voice_es_19704715.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1398 + }, + { + "path": "common_voice_en_19724779.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1399 + }, + { + "path": "common_voice_en_18937866.mp3", + "question": "The speaker is using which language?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1400 + }, + { + "path": "common_voice_zh-CN_18997095.mp3", + "question": "The speech was in which language?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1401 + }, + { + "path": "common_voice_en_19382165.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1402 + }, + { + "path": "common_voice_it_19990637.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1403 + }, + { + "path": "common_voice_zh-CN_20070843.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1404 + }, + { + "path": "common_voice_en_138829.mp3", + "question": "identify the language of this speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1405 + }, + { + "path": "common_voice_es_19637780.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1406 + }, + { + "path": "common_voice_es_19618278.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1407 + }, + { + "path": "common_voice_en_19735775.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1408 + }, + { + "path": "common_voice_es_19627628.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1409 + }, + { + "path": "common_voice_zh-CN_18987390.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1410 + }, + { + "path": "common_voice_en_19942747.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1411 + }, + { + "path": "common_voice_en_2544322.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1412 + }, + { + "path": "common_voice_en_20054204.mp3", + "question": "Which language is used in this speech?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1413 + }, + { + "path": "common_voice_en_14888.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1414 + }, + { + "path": "common_voice_it_19991121.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1415 + }, + { + "path": "common_voice_en_19651160.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1416 + }, + { + "path": "common_voice_en_678305.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1417 + }, + { + "path": "common_voice_fr_17628160.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1418 + }, + { + "path": "common_voice_es_19596879.mp3", + "question": "Identify the language of the speech.", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1419 + }, + { + "path": "common_voice_es_19643409.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1420 + }, + { + "path": "common_voice_it_17985022.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1421 + }, + { + "path": "common_voice_fr_19793960.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1422 + }, + { + "path": "common_voice_en_19684739.mp3", + "question": "Which language is used in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1423 + }, + { + "path": "common_voice_en_20008888.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1424 + }, + { + "path": "common_voice_en_19947732.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1425 + }, + { + "path": "common_voice_en_561984.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1426 + }, + { + "path": "common_voice_it_20011647.mp3", + "question": "What language is the speech in?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1427 + }, + { + "path": "common_voice_en_20050823.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1428 + }, + { + "path": "common_voice_en_18474112.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1429 + }, + { + "path": "common_voice_fr_19679173.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1430 + }, + { + "path": "common_voice_es_19609854.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1431 + }, + { + "path": "common_voice_en_18456605.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1432 + }, + { + "path": "common_voice_es_19769296.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1433 + }, + { + "path": "common_voice_en_19677497.mp3", + "question": "In which language is this speech being given?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1434 + }, + { + "path": "common_voice_en_17874676.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1435 + }, + { + "path": "common_voice_zh-CN_18979337.mp3", + "question": "The speech was presented in what language?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1436 + }, + { + "path": "common_voice_fr_19616322.mp3", + "question": "Identify the language from the speech.", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1437 + }, + { + "path": "common_voice_en_19667814.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1438 + }, + { + "path": "common_voice_es_19725515.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1439 + }, + { + "path": "common_voice_de_19167141.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1440 + }, + { + "path": "common_voice_fr_19717792.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1441 + }, + { + "path": "common_voice_es_19636530.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1442 + }, + { + "path": "common_voice_de_17407683.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1443 + }, + { + "path": "common_voice_en_18468174.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1444 + }, + { + "path": "common_voice_it_19971540.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1445 + }, + { + "path": "common_voice_de_19151304.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1446 + }, + { + "path": "common_voice_es_19764184.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1447 + }, + { + "path": "common_voice_zh-CN_18990594.mp3", + "question": "Guess the language of the speech.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1448 + }, + { + "path": "common_voice_en_19277729.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1449 + }, + { + "path": "common_voice_en_17284403.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1450 + }, + { + "path": "common_voice_de_19162145.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1451 + }, + { + "path": "common_voice_de_19158794.mp3", + "question": "In which language is this speech being given?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1452 + }, + { + "path": "common_voice_fr_19707295.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1453 + }, + { + "path": "common_voice_it_17832713.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1454 + }, + { + "path": "common_voice_fr_17393476.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1455 + }, + { + "path": "common_voice_zh-CN_19427085.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1456 + }, + { + "path": "common_voice_en_18320249.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1457 + }, + { + "path": "common_voice_en_18475045.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1458 + }, + { + "path": "common_voice_en_20011635.mp3", + "question": "Identify the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1459 + }, + { + "path": "common_voice_en_19729647.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1460 + }, + { + "path": "common_voice_fr_19754548.mp3", + "question": "What language is the speech in?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1461 + }, + { + "path": "common_voice_en_19768052.mp3", + "question": "Guess the language in the audio.", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1462 + }, + { + "path": "common_voice_en_18345757.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1463 + }, + { + "path": "common_voice_de_20005545.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1464 + }, + { + "path": "common_voice_fr_18108946.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1465 + }, + { + "path": "common_voice_zh-CN_18810213.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1466 + }, + { + "path": "common_voice_zh-CN_18654411.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1467 + }, + { + "path": "common_voice_fr_19001230.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1468 + }, + { + "path": "common_voice_en_19732028.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1469 + }, + { + "path": "common_voice_en_19614570.mp3", + "question": "What language is used in this speech?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1470 + }, + { + "path": "common_voice_fr_18926587.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1471 + }, + { + "path": "common_voice_es_19996229.mp3", + "question": "The speech was in which language?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1472 + }, + { + "path": "common_voice_es_19956851.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1473 + }, + { + "path": "common_voice_en_633686.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1474 + }, + { + "path": "common_voice_en_19747280.mp3", + "question": "What language is used in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1475 + }, + { + "path": "common_voice_fr_19682094.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1476 + }, + { + "path": "common_voice_es_19760059.mp3", + "question": "In which language was the announcement made?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1477 + }, + { + "path": "common_voice_en_573548.mp3", + "question": "The speech was in which language?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1478 + }, + { + "path": "common_voice_it_17822332.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1479 + }, + { + "path": "common_voice_fr_18474830.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1480 + }, + { + "path": "common_voice_de_17333733.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1481 + }, + { + "path": "common_voice_fr_19689849.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1482 + }, + { + "path": "common_voice_fr_19971792.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1483 + }, + { + "path": "common_voice_it_19993255.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1484 + }, + { + "path": "common_voice_en_479942.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1485 + }, + { + "path": "common_voice_fr_19757867.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1486 + }, + { + "path": "common_voice_es_19134264.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1487 + }, + { + "path": "common_voice_it_17540729.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1488 + }, + { + "path": "common_voice_en_19710738.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1489 + }, + { + "path": "common_voice_en_19713613.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1490 + }, + { + "path": "common_voice_fr_19605372.mp3", + "question": "What language is used in this speech?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1491 + }, + { + "path": "common_voice_en_19769647.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1492 + }, + { + "path": "common_voice_es_19626432.mp3", + "question": "The speech was in which language?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1493 + }, + { + "path": "common_voice_fr_19676707.mp3", + "question": "What language is the speaker using?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1494 + }, + { + "path": "common_voice_zh-CN_19491053.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1495 + }, + { + "path": "common_voice_de_19161779.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1496 + }, + { + "path": "common_voice_en_533152.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1497 + }, + { + "path": "common_voice_en_67655.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1498 + }, + { + "path": "common_voice_en_20067157.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1499 + }, + { + "path": "common_voice_fr_18340694.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1500 + }, + { + "path": "common_voice_es_19610053.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1501 + }, + { + "path": "common_voice_en_19953570.mp3", + "question": "The speaker is using which language?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1502 + }, + { + "path": "common_voice_en_20045416.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1503 + }, + { + "path": "common_voice_de_17871735.mp3", + "question": "In which language is this speech being given?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1504 + }, + { + "path": "common_voice_en_18963300.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1505 + }, + { + "path": "common_voice_it_20004287.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1506 + }, + { + "path": "common_voice_zh-CN_18776489.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1507 + }, + { + "path": "common_voice_en_18988530.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1508 + }, + { + "path": "common_voice_en_19622628.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1509 + }, + { + "path": "common_voice_zh-CN_18978209.mp3", + "question": "Guess the language of the speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1510 + }, + { + "path": "common_voice_it_20035903.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1511 + }, + { + "path": "common_voice_es_19616052.mp3", + "question": "Which language is used in this speech?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1512 + }, + { + "path": "common_voice_en_187743.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1513 + }, + { + "path": "common_voice_en_18839531.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1514 + }, + { + "path": "common_voice_en_180176.mp3", + "question": "What language is used in this speech?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1515 + }, + { + "path": "common_voice_de_19124128.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1516 + }, + { + "path": "common_voice_fr_19737844.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1517 + }, + { + "path": "common_voice_fr_19631949.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1518 + }, + { + "path": "common_voice_es_19654078.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1519 + }, + { + "path": "common_voice_fr_19999068.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1520 + }, + { + "path": "common_voice_en_520490.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1521 + }, + { + "path": "common_voice_it_20019591.mp3", + "question": "identify the language of this speech.", + "choice_a": "it", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1522 + }, + { + "path": "common_voice_en_19598193.mp3", + "question": "In which language was the announcement made?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1523 + }, + { + "path": "common_voice_fr_17320164.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1524 + }, + { + "path": "common_voice_fr_20044070.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1525 + }, + { + "path": "common_voice_en_193221.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1526 + }, + { + "path": "common_voice_en_150709.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1527 + }, + { + "path": "common_voice_it_18655348.mp3", + "question": "What language is the speaker using?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1528 + }, + { + "path": "common_voice_es_19120964.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1529 + }, + { + "path": "common_voice_en_533211.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1530 + }, + { + "path": "common_voice_en_75208.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1531 + }, + { + "path": "common_voice_it_20041416.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1532 + }, + { + "path": "common_voice_fr_19641516.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1533 + }, + { + "path": "common_voice_fr_17913304.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1534 + }, + { + "path": "common_voice_fr_19602260.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1535 + }, + { + "path": "common_voice_zh-CN_18982719.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1536 + }, + { + "path": "common_voice_en_17292253.mp3", + "question": "In which language was the announcement made?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1537 + }, + { + "path": "common_voice_it_20016882.mp3", + "question": "identify the language of this speech.", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1538 + }, + { + "path": "common_voice_en_19983000.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1539 + }, + { + "path": "common_voice_fr_19767530.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1540 + }, + { + "path": "common_voice_it_18151690.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1541 + }, + { + "path": "common_voice_en_19750358.mp3", + "question": "The speech was presented in what language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1542 + }, + { + "path": "common_voice_de_19170281.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1543 + }, + { + "path": "common_voice_zh-CN_18597877.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1544 + }, + { + "path": "common_voice_en_19620661.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1545 + }, + { + "path": "common_voice_en_19758543.mp3", + "question": "In which language is this speech being given?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1546 + }, + { + "path": "common_voice_fr_19735869.mp3", + "question": "In which language was the announcement made?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1547 + }, + { + "path": "common_voice_en_19768725.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1548 + }, + { + "path": "common_voice_en_19611598.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1549 + }, + { + "path": "common_voice_fr_20021470.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1550 + }, + { + "path": "common_voice_en_17796149.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1551 + }, + { + "path": "common_voice_en_18812064.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1552 + }, + { + "path": "common_voice_es_19625034.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "es", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1553 + }, + { + "path": "common_voice_fr_19506756.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1554 + }, + { + "path": "common_voice_en_19689281.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1555 + }, + { + "path": "common_voice_en_1405134.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1556 + }, + { + "path": "common_voice_de_18549986.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1557 + }, + { + "path": "common_voice_en_19964439.mp3", + "question": "Guess the language in the audio.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1558 + }, + { + "path": "common_voice_de_17870687.mp3", + "question": "In which language was the announcement made?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1559 + }, + { + "path": "common_voice_es_19956591.mp3", + "question": "Guess the language of the speech.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1560 + }, + { + "path": "common_voice_de_19682858.mp3", + "question": "Guess the language in the audio.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1561 + }, + { + "path": "common_voice_de_19964836.mp3", + "question": "The speech was in which language?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1562 + }, + { + "path": "common_voice_zh-CN_18982711.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1563 + }, + { + "path": "common_voice_en_18936099.mp3", + "question": "Identify the language from the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1564 + }, + { + "path": "common_voice_fr_19966648.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1565 + }, + { + "path": "common_voice_zh-CN_18978638.mp3", + "question": "The speech was in which language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1566 + }, + { + "path": "common_voice_en_20033246.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1567 + }, + { + "path": "common_voice_en_19959625.mp3", + "question": "The speech was presented in what language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1568 + }, + { + "path": "common_voice_de_18854869.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1569 + }, + { + "path": "common_voice_en_17871648.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1570 + }, + { + "path": "common_voice_fr_18965992.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1571 + }, + { + "path": "common_voice_es_19613562.mp3", + "question": "The speaker is using which language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1572 + }, + { + "path": "common_voice_de_17959028.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1573 + }, + { + "path": "common_voice_en_579986.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1574 + }, + { + "path": "common_voice_es_19028498.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1575 + }, + { + "path": "common_voice_it_18854994.mp3", + "question": "What language is the speech in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1576 + }, + { + "path": "common_voice_en_18960792.mp3", + "question": "What language is used in this speech?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1577 + }, + { + "path": "common_voice_de_19190383.mp3", + "question": "Which language is used in this speech?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1578 + }, + { + "path": "common_voice_en_18949953.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1579 + }, + { + "path": "common_voice_en_19562676.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1580 + }, + { + "path": "common_voice_en_19867612.mp3", + "question": "What language is the speaker using?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1581 + }, + { + "path": "common_voice_en_18845658.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1582 + }, + { + "path": "common_voice_en_18603199.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1583 + }, + { + "path": "common_voice_en_20089584.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1584 + }, + { + "path": "common_voice_fr_17815253.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1585 + }, + { + "path": "common_voice_en_19984962.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1586 + }, + { + "path": "common_voice_es_19130136.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1587 + }, + { + "path": "common_voice_en_18466091.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1588 + }, + { + "path": "common_voice_fr_17920906.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1589 + }, + { + "path": "common_voice_de_18049993.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1590 + }, + { + "path": "common_voice_en_575266.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1591 + }, + { + "path": "common_voice_en_17895867.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1592 + }, + { + "path": "common_voice_en_19578323.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1593 + }, + { + "path": "common_voice_it_19459543.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1594 + }, + { + "path": "common_voice_fr_18485488.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1595 + }, + { + "path": "common_voice_en_641881.mp3", + "question": "What language is used in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1596 + }, + { + "path": "common_voice_fr_18058990.mp3", + "question": "Identify the language from the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1597 + }, + { + "path": "common_voice_fr_19680574.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1598 + }, + { + "path": "common_voice_de_19170565.mp3", + "question": "Guess the language in the audio.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1599 + }, + { + "path": "common_voice_en_18039748.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1600 + }, + { + "path": "common_voice_en_18414126.mp3", + "question": "The speech was presented in what language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1601 + }, + { + "path": "common_voice_zh-CN_19526098.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1602 + }, + { + "path": "common_voice_de_19983784.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1603 + }, + { + "path": "common_voice_de_18339635.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1604 + }, + { + "path": "common_voice_en_660250.mp3", + "question": "Guess the language in the audio.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1605 + }, + { + "path": "common_voice_es_19723340.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1606 + }, + { + "path": "common_voice_it_20059401.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1607 + }, + { + "path": "common_voice_it_19970024.mp3", + "question": "In which language was the announcement made?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1608 + }, + { + "path": "common_voice_en_19699263.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1609 + }, + { + "path": "common_voice_en_19676676.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1610 + }, + { + "path": "common_voice_de_19262400.mp3", + "question": "In which language was the announcement made?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1611 + }, + { + "path": "common_voice_en_390876.mp3", + "question": "Which language is used in this speech?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1612 + }, + { + "path": "common_voice_en_18902775.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1613 + }, + { + "path": "common_voice_en_17715740.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1614 + }, + { + "path": "common_voice_en_18733700.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1615 + }, + { + "path": "common_voice_it_19989940.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1616 + }, + { + "path": "common_voice_it_19993395.mp3", + "question": "Identify the language of the speech.", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1617 + }, + { + "path": "common_voice_en_19951091.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1618 + }, + { + "path": "common_voice_en_19972117.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1619 + }, + { + "path": "common_voice_en_18066344.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1620 + }, + { + "path": "common_voice_en_19945480.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1621 + }, + { + "path": "common_voice_en_18302125.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1622 + }, + { + "path": "common_voice_en_19886147.mp3", + "question": "Identify the language from the speech.", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1623 + }, + { + "path": "common_voice_en_19539191.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1624 + }, + { + "path": "common_voice_en_158505.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1625 + }, + { + "path": "common_voice_en_216142.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1626 + }, + { + "path": "common_voice_en_65535.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1627 + }, + { + "path": "common_voice_es_19697775.mp3", + "question": "What language is the speech in?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1628 + }, + { + "path": "common_voice_fr_19955278.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1629 + }, + { + "path": "common_voice_es_18538439.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1630 + }, + { + "path": "common_voice_en_19603833.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1631 + }, + { + "path": "common_voice_en_586565.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1632 + }, + { + "path": "common_voice_en_579853.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1633 + }, + { + "path": "common_voice_es_19596988.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1634 + }, + { + "path": "common_voice_es_19762492.mp3", + "question": "The speech was in which language?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1635 + }, + { + "path": "common_voice_en_19605840.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1636 + }, + { + "path": "common_voice_es_20060371.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1637 + }, + { + "path": "common_voice_en_19601617.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1638 + }, + { + "path": "common_voice_fr_19679903.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1639 + }, + { + "path": "common_voice_en_18965046.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1640 + }, + { + "path": "common_voice_es_19975442.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1641 + }, + { + "path": "common_voice_it_19977717.mp3", + "question": "The speech was presented in what language?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1642 + }, + { + "path": "common_voice_fr_17352671.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1643 + }, + { + "path": "common_voice_de_19704152.mp3", + "question": "What language is the speech in?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1644 + }, + { + "path": "common_voice_en_48069.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1645 + }, + { + "path": "common_voice_it_19993401.mp3", + "question": "The speaker is using which language?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1646 + }, + { + "path": "common_voice_en_138777.mp3", + "question": "Guess the language in the audio.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1647 + }, + { + "path": "common_voice_de_19092931.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1648 + }, + { + "path": "common_voice_es_19606807.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1649 + }, + { + "path": "common_voice_zh-CN_18729661.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1650 + }, + { + "path": "common_voice_es_18727615.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1651 + }, + { + "path": "common_voice_es_19804651.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1652 + }, + { + "path": "common_voice_en_105881.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1653 + }, + { + "path": "common_voice_en_478570.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1654 + }, + { + "path": "common_voice_fr_17410059.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1655 + }, + { + "path": "common_voice_en_18733288.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1656 + }, + { + "path": "common_voice_en_6560861.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1657 + }, + { + "path": "common_voice_en_19625640.mp3", + "question": "Identify the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1658 + }, + { + "path": "common_voice_es_19980341.mp3", + "question": "Guess the language of the speech.", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1659 + }, + { + "path": "common_voice_en_62618.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1660 + }, + { + "path": "common_voice_es_20026965.mp3", + "question": "What language is used in this speech?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1661 + }, + { + "path": "common_voice_en_19974321.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1662 + }, + { + "path": "common_voice_es_19611046.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1663 + }, + { + "path": "common_voice_fr_19675271.mp3", + "question": "The speech was presented in what language?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1664 + }, + { + "path": "common_voice_de_19730369.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1665 + }, + { + "path": "common_voice_en_63837.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1666 + }, + { + "path": "common_voice_en_19887361.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1667 + }, + { + "path": "common_voice_de_19631065.mp3", + "question": "The speech was in which language?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1668 + }, + { + "path": "common_voice_fr_19631503.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1669 + }, + { + "path": "common_voice_en_19684654.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1670 + }, + { + "path": "common_voice_en_19950595.mp3", + "question": "Guess the language of the speech.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1671 + }, + { + "path": "common_voice_en_19546230.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1672 + }, + { + "path": "common_voice_en_15734651.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1673 + }, + { + "path": "common_voice_en_18304350.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1674 + }, + { + "path": "common_voice_en_20027505.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1675 + }, + { + "path": "common_voice_en_19606478.mp3", + "question": "In which language is this speech being given?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1676 + }, + { + "path": "common_voice_en_71422.mp3", + "question": "Guess the language of the speech.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1677 + }, + { + "path": "common_voice_de_19719923.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1678 + }, + { + "path": "common_voice_en_19700152.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1679 + }, + { + "path": "common_voice_de_19731119.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1680 + }, + { + "path": "common_voice_en_17278773.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1681 + }, + { + "path": "common_voice_en_17934330.mp3", + "question": "Guess the language in the audio.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1682 + }, + { + "path": "common_voice_it_17788449.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1683 + }, + { + "path": "common_voice_en_19956221.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1684 + }, + { + "path": "common_voice_es_19714364.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1685 + }, + { + "path": "common_voice_en_19680644.mp3", + "question": "Guess the language in the audio.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1686 + }, + { + "path": "common_voice_fr_19631568.mp3", + "question": "The speech was in which language?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1687 + }, + { + "path": "common_voice_en_19974640.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1688 + }, + { + "path": "common_voice_es_19599589.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1689 + }, + { + "path": "common_voice_zh-CN_18663187.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1690 + }, + { + "path": "common_voice_de_19116856.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1691 + }, + { + "path": "common_voice_en_18737799.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1692 + }, + { + "path": "common_voice_it_19985020.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1693 + }, + { + "path": "common_voice_en_19869672.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1694 + }, + { + "path": "common_voice_en_20047926.mp3", + "question": "Identify the language from the speech.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1695 + }, + { + "path": "common_voice_en_20113323.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1696 + }, + { + "path": "common_voice_de_19949337.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1697 + }, + { + "path": "common_voice_de_19389673.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1698 + }, + { + "path": "common_voice_es_19991757.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1699 + }, + { + "path": "common_voice_de_20003692.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1700 + }, + { + "path": "common_voice_de_19152919.mp3", + "question": "The speech was presented in what language?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1701 + }, + { + "path": "common_voice_en_19358679.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1702 + }, + { + "path": "common_voice_es_19610925.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1703 + }, + { + "path": "common_voice_en_18641793.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1704 + }, + { + "path": "common_voice_es_20025124.mp3", + "question": "What language is used in this speech?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1705 + }, + { + "path": "common_voice_en_18771748.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1706 + }, + { + "path": "common_voice_en_117433.mp3", + "question": "The speaker is using which language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1707 + }, + { + "path": "common_voice_es_18519105.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1708 + }, + { + "path": "common_voice_en_19465314.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1709 + }, + { + "path": "common_voice_it_18543264.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1710 + }, + { + "path": "common_voice_en_20065295.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1711 + }, + { + "path": "common_voice_de_19997462.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1712 + }, + { + "path": "common_voice_en_19370209.mp3", + "question": "The speech was presented in what language?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1713 + }, + { + "path": "common_voice_en_19701366.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1714 + }, + { + "path": "common_voice_it_20005940.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1715 + }, + { + "path": "common_voice_de_18195774.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1716 + }, + { + "path": "common_voice_de_18435951.mp3", + "question": "Guess the language in the audio.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1717 + }, + { + "path": "common_voice_en_18509191.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1718 + }, + { + "path": "common_voice_en_18270076.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1719 + }, + { + "path": "common_voice_en_521963.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1720 + }, + { + "path": "common_voice_en_19745289.mp3", + "question": "Guess the language in the audio.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1721 + }, + { + "path": "common_voice_de_19965495.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1722 + }, + { + "path": "common_voice_fr_19504245.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1723 + }, + { + "path": "common_voice_en_188339.mp3", + "question": "Identify the language from the speech.", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1724 + }, + { + "path": "common_voice_de_19422939.mp3", + "question": "Identify the language from the speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1725 + }, + { + "path": "common_voice_de_19952744.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1726 + }, + { + "path": "common_voice_en_19758089.mp3", + "question": "What language is the speaker using?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1727 + }, + { + "path": "common_voice_es_19651183.mp3", + "question": "Identify the language from the speech.", + "choice_a": "es", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1728 + }, + { + "path": "common_voice_en_19947773.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1729 + }, + { + "path": "common_voice_en_585935.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1730 + }, + { + "path": "common_voice_en_19842786.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1731 + }, + { + "path": "common_voice_en_18909486.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1732 + }, + { + "path": "common_voice_it_17444561.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1733 + }, + { + "path": "common_voice_en_490358.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1734 + }, + { + "path": "common_voice_de_18773540.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1735 + }, + { + "path": "common_voice_en_174349.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1736 + }, + { + "path": "common_voice_es_19760401.mp3", + "question": "Guess the language in the audio.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1737 + }, + { + "path": "common_voice_zh-CN_18659534.mp3", + "question": "The speech was in which language?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1738 + }, + { + "path": "common_voice_en_19729024.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1739 + }, + { + "path": "common_voice_en_613811.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1740 + }, + { + "path": "common_voice_de_17303907.mp3", + "question": "identify the language of this speech.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1741 + }, + { + "path": "common_voice_en_544916.mp3", + "question": "What language is used in this speech?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1742 + }, + { + "path": "common_voice_es_19613835.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1743 + }, + { + "path": "common_voice_en_18273546.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1744 + }, + { + "path": "common_voice_en_19818233.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1745 + }, + { + "path": "common_voice_en_18936020.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1746 + }, + { + "path": "common_voice_fr_19647806.mp3", + "question": "What language is used in this speech?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1747 + }, + { + "path": "common_voice_en_18504873.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1748 + }, + { + "path": "common_voice_de_19449503.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1749 + }, + { + "path": "common_voice_es_19614394.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1750 + }, + { + "path": "common_voice_en_19776787.mp3", + "question": "Which language is used in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1751 + }, + { + "path": "common_voice_en_19617093.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1752 + }, + { + "path": "common_voice_fr_19386263.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1753 + }, + { + "path": "common_voice_es_19725513.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1754 + }, + { + "path": "common_voice_en_19675621.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "fr", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1755 + }, + { + "path": "common_voice_it_19989540.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1756 + }, + { + "path": "common_voice_en_1416457.mp3", + "question": "The speech was presented in what language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1757 + }, + { + "path": "common_voice_en_32938.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1758 + }, + { + "path": "common_voice_de_18965694.mp3", + "question": "What language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1759 + }, + { + "path": "common_voice_en_20048755.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1760 + }, + { + "path": "common_voice_de_18217473.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1761 + }, + { + "path": "common_voice_fr_19651629.mp3", + "question": "In which language was the announcement made?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1762 + }, + { + "path": "common_voice_it_19990132.mp3", + "question": "Guess the language of the speech.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1763 + }, + { + "path": "common_voice_en_19984546.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1764 + }, + { + "path": "common_voice_fr_17790381.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1765 + }, + { + "path": "common_voice_en_19456202.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1766 + }, + { + "path": "common_voice_en_518763.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1767 + }, + { + "path": "common_voice_fr_19659850.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1768 + }, + { + "path": "common_voice_de_19774430.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1769 + }, + { + "path": "common_voice_en_20023864.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1770 + }, + { + "path": "common_voice_en_19994962.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1771 + }, + { + "path": "common_voice_it_20041068.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1772 + }, + { + "path": "common_voice_en_19160955.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1773 + }, + { + "path": "common_voice_de_19847029.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1774 + }, + { + "path": "common_voice_en_20046674.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1775 + }, + { + "path": "common_voice_en_934751.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1776 + }, + { + "path": "common_voice_zh-CN_19015125.mp3", + "question": "Identify the language of the speech.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1777 + }, + { + "path": "common_voice_en_19615979.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1778 + }, + { + "path": "common_voice_fr_19737998.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1779 + }, + { + "path": "common_voice_en_19606157.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1780 + }, + { + "path": "common_voice_en_19439978.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1781 + }, + { + "path": "common_voice_en_20048329.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1782 + }, + { + "path": "common_voice_zh-CN_19186158.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1783 + }, + { + "path": "common_voice_es_18971344.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1784 + }, + { + "path": "common_voice_fr_17312931.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1785 + }, + { + "path": "common_voice_en_17856800.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1786 + }, + { + "path": "common_voice_en_18415727.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1787 + }, + { + "path": "common_voice_es_19767990.mp3", + "question": "Guess the language in the audio.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1788 + }, + { + "path": "common_voice_fr_19606011.mp3", + "question": "The speech was in which language?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1789 + }, + { + "path": "common_voice_en_19942971.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1790 + }, + { + "path": "common_voice_es_19655827.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1791 + }, + { + "path": "common_voice_de_20011693.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "it", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1792 + }, + { + "path": "common_voice_en_18592565.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1793 + }, + { + "path": "common_voice_es_19673826.mp3", + "question": "Identify the language of the speech.", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1794 + }, + { + "path": "common_voice_en_19417340.mp3", + "question": "The speaker is using which language?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1795 + }, + { + "path": "common_voice_en_19629784.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1796 + }, + { + "path": "common_voice_de_17815147.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1797 + }, + { + "path": "common_voice_fr_17393025.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1798 + }, + { + "path": "common_voice_de_19020996.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1799 + }, + { + "path": "common_voice_de_19417385.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1800 + }, + { + "path": "common_voice_zh-CN_18844480.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1801 + }, + { + "path": "common_voice_en_19682749.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1802 + }, + { + "path": "common_voice_zh-CN_18983164.mp3", + "question": "What is the language of the speaker in the audio clip?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1803 + }, + { + "path": "common_voice_it_19973172.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1804 + }, + { + "path": "common_voice_en_18451362.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1805 + }, + { + "path": "common_voice_fr_19734937.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1806 + }, + { + "path": "common_voice_es_19655027.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1807 + }, + { + "path": "common_voice_en_19674586.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1808 + }, + { + "path": "common_voice_en_19637442.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1809 + }, + { + "path": "common_voice_fr_19642686.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1810 + }, + { + "path": "common_voice_en_17296199.mp3", + "question": "What language is the speaker using?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1811 + }, + { + "path": "common_voice_de_17989759.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1812 + }, + { + "path": "common_voice_fr_19966435.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1813 + }, + { + "path": "common_voice_en_19757970.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "it", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1814 + }, + { + "path": "common_voice_en_17883942.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1815 + }, + { + "path": "common_voice_de_18868756.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1816 + }, + { + "path": "common_voice_en_18848261.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1817 + }, + { + "path": "common_voice_zh-CN_19773551.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1818 + }, + { + "path": "common_voice_en_13963687.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1819 + }, + { + "path": "common_voice_en_20011157.mp3", + "question": "Which language is used in this speech?", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1820 + }, + { + "path": "common_voice_en_19598680.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1821 + }, + { + "path": "common_voice_es_19626892.mp3", + "question": "In which language is this speech being given?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1822 + }, + { + "path": "common_voice_en_537389.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1823 + }, + { + "path": "common_voice_es_19659027.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1824 + }, + { + "path": "common_voice_en_19961967.mp3", + "question": "Identify the language from the speech excerpt.", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1825 + }, + { + "path": "common_voice_en_130021.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1826 + }, + { + "path": "common_voice_it_20030410.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1827 + }, + { + "path": "common_voice_fr_17744882.mp3", + "question": "Which language is used in this speech?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1828 + }, + { + "path": "common_voice_es_19659163.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1829 + }, + { + "path": "common_voice_de_19757350.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1830 + }, + { + "path": "common_voice_de_19701664.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1831 + }, + { + "path": "common_voice_zh-CN_18776316.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1832 + }, + { + "path": "common_voice_en_18538098.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1833 + }, + { + "path": "common_voice_zh-CN_18992306.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1834 + }, + { + "path": "common_voice_en_15735718.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1835 + }, + { + "path": "common_voice_en_31993.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1836 + }, + { + "path": "common_voice_en_110087.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1837 + }, + { + "path": "common_voice_en_19751294.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1838 + }, + { + "path": "common_voice_en_18468175.mp3", + "question": "What language is the speaker using?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1839 + }, + { + "path": "common_voice_en_244262.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1840 + }, + { + "path": "common_voice_es_19622578.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1841 + }, + { + "path": "common_voice_en_19735675.mp3", + "question": "The speech was in which language?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1842 + }, + { + "path": "common_voice_en_599910.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1843 + }, + { + "path": "common_voice_en_19968214.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1844 + }, + { + "path": "common_voice_en_19651272.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1845 + }, + { + "path": "common_voice_en_18733709.mp3", + "question": "Identify the language from the speech.", + "choice_a": "en", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1846 + }, + { + "path": "common_voice_it_17700750.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1847 + }, + { + "path": "common_voice_en_18095211.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1848 + }, + { + "path": "common_voice_it_20036767.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1849 + }, + { + "path": "common_voice_en_338044.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1850 + }, + { + "path": "common_voice_zh-CN_18986638.mp3", + "question": "Determine the language used in this speech.", + "choice_a": "es", + "choice_b": "en", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1851 + }, + { + "path": "common_voice_fr_19502631.mp3", + "question": "Which language is the speaker speaking in?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1852 + }, + { + "path": "common_voice_de_18844542.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1853 + }, + { + "path": "common_voice_en_579908.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1854 + }, + { + "path": "common_voice_de_19645702.mp3", + "question": "What language is the speech in?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1855 + }, + { + "path": "common_voice_en_19639638.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1856 + }, + { + "path": "common_voice_es_19139957.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1857 + }, + { + "path": "common_voice_de_18445003.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1858 + }, + { + "path": "common_voice_en_454129.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1859 + }, + { + "path": "common_voice_en_4207119.mp3", + "question": "What language is the speech in?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1860 + }, + { + "path": "common_voice_en_19754169.mp3", + "question": "What language is the speaker using?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1861 + }, + { + "path": "common_voice_en_129890.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "fr", + "choice_b": "it", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1862 + }, + { + "path": "common_voice_zh-CN_18552866.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "de", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1863 + }, + { + "path": "common_voice_en_19739839.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1864 + }, + { + "path": "common_voice_en_75901.mp3", + "question": "Guess the language of the speech.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1865 + }, + { + "path": "common_voice_fr_19645973.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1866 + }, + { + "path": "common_voice_en_152341.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1867 + }, + { + "path": "common_voice_es_20062044.mp3", + "question": "Identify the language of the speech.", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1868 + }, + { + "path": "common_voice_fr_19140964.mp3", + "question": "The speech was presented in what language?", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1869 + }, + { + "path": "common_voice_it_17939126.mp3", + "question": "Which language is being delivered in this speech?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1870 + }, + { + "path": "common_voice_de_20002685.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "zh-CN", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1871 + }, + { + "path": "common_voice_fr_20005591.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1872 + }, + { + "path": "common_voice_en_19597842.mp3", + "question": "What language is the speech in?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1873 + }, + { + "path": "common_voice_de_19549354.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1874 + }, + { + "path": "common_voice_en_19612287.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1875 + }, + { + "path": "common_voice_it_19997445.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1876 + }, + { + "path": "common_voice_es_19980897.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1877 + }, + { + "path": "common_voice_en_19678886.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1878 + }, + { + "path": "common_voice_en_41885.mp3", + "question": "Which language is used in this speech?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1879 + }, + { + "path": "common_voice_en_18517443.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1880 + }, + { + "path": "common_voice_fr_19625388.mp3", + "question": "Tell which language this speech is narrated in?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1881 + }, + { + "path": "common_voice_fr_19626330.mp3", + "question": "Which language can you hear in this speech?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1882 + }, + { + "path": "common_voice_es_19696383.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1883 + }, + { + "path": "common_voice_de_19031350.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "es", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1884 + }, + { + "path": "common_voice_es_19699586.mp3", + "question": "Which language is used in this speech?", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1885 + }, + { + "path": "common_voice_es_19617025.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1886 + }, + { + "path": "common_voice_it_20023930.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1887 + }, + { + "path": "common_voice_es_19955867.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1888 + }, + { + "path": "common_voice_en_17965262.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1889 + }, + { + "path": "common_voice_de_19661732.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "es", + "choice_b": "fr", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1890 + }, + { + "path": "common_voice_zh-CN_18979263.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "ja", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1891 + }, + { + "path": "common_voice_en_20149508.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1892 + }, + { + "path": "common_voice_zh-CN_19511890.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "it", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1893 + }, + { + "path": "common_voice_en_20023054.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1894 + }, + { + "path": "common_voice_es_19720180.mp3", + "question": "In which language is this speech being given?", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1895 + }, + { + "path": "common_voice_en_13188.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1896 + }, + { + "path": "common_voice_en_18171167.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1897 + }, + { + "path": "common_voice_fr_19594057.mp3", + "question": "In which language was the announcement made?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1898 + }, + { + "path": "common_voice_zh-CN_18978869.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1899 + }, + { + "path": "common_voice_it_19972590.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1900 + }, + { + "path": "common_voice_en_18851501.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1901 + }, + { + "path": "common_voice_en_19696977.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1902 + }, + { + "path": "common_voice_fr_17969282.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1903 + }, + { + "path": "common_voice_fr_19193564.mp3", + "question": "Identify the language from the speech.", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1904 + }, + { + "path": "common_voice_en_19944700.mp3", + "question": "In which language is the speaker presenting?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1905 + }, + { + "path": "common_voice_en_19684360.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1906 + }, + { + "path": "common_voice_en_20016363.mp3", + "question": "Identify the language of the speech.", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1907 + }, + { + "path": "common_voice_es_19659396.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1908 + }, + { + "path": "common_voice_en_19856737.mp3", + "question": "What is the official language of the speech you just heard?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1909 + }, + { + "path": "common_voice_en_19753938.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1910 + }, + { + "path": "common_voice_en_18485471.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1911 + }, + { + "path": "common_voice_de_18202305.mp3", + "question": "Which language is being spoken in the audio?", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1912 + }, + { + "path": "common_voice_es_19130353.mp3", + "question": "In which language is this speech being given?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1913 + }, + { + "path": "common_voice_en_25299.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "fr", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1914 + }, + { + "path": "common_voice_it_20003082.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "it", + "choice_b": "en", + "choice_c": "de", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1915 + }, + { + "path": "common_voice_fr_19009861.mp3", + "question": "In which language was the announcement made?", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1916 + }, + { + "path": "common_voice_en_177683.mp3", + "question": "Recognize the language of the speech.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1917 + }, + { + "path": "common_voice_es_18618158.mp3", + "question": "In which language was the announcement made?", + "choice_a": "es", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1918 + }, + { + "path": "common_voice_fr_19761366.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1919 + }, + { + "path": "common_voice_fr_18655587.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "zh-CN", + "choice_b": "fr", + "choice_c": "en", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1920 + }, + { + "path": "common_voice_it_19979306.mp3", + "question": "Identify the language from the speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1921 + }, + { + "path": "common_voice_de_19120207.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "es", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1922 + }, + { + "path": "common_voice_it_17705450.mp3", + "question": "In which language was the announcement made?", + "choice_a": "de", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1923 + }, + { + "path": "common_voice_en_18782394.mp3", + "question": "Guess the language of the speech.", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1924 + }, + { + "path": "common_voice_en_15904536.mp3", + "question": "Determine the language of the dialogue in this snippet.", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1925 + }, + { + "path": "common_voice_es_19839394.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1926 + }, + { + "path": "common_voice_en_19257110.mp3", + "question": "Identify the language of the conversation you just heard.", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1927 + }, + { + "path": "common_voice_en_560935.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "ja", + "choice_b": "it", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1928 + }, + { + "path": "common_voice_es_19942405.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "it", + "choice_b": "en", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1929 + }, + { + "path": "common_voice_en_542262.mp3", + "question": "Identify the language from the speech.", + "choice_a": "it", + "choice_b": "de", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1930 + }, + { + "path": "common_voice_de_19119577.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "it", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1931 + }, + { + "path": "common_voice_zh-CN_18978562.mp3", + "question": "Identify the language of the speech.", + "choice_a": "en", + "choice_b": "de", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1932 + }, + { + "path": "common_voice_en_19746685.mp3", + "question": "The speech was in which language?", + "choice_a": "ja", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1933 + }, + { + "path": "common_voice_it_20032384.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1934 + }, + { + "path": "common_voice_es_19980861.mp3", + "question": "The speech was in which language?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1935 + }, + { + "path": "common_voice_en_17726300.mp3", + "question": "What language is being used for the speaker in this speech?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "zh-CN", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1936 + }, + { + "path": "common_voice_en_19699680.mp3", + "question": "Identify the language of the speaker.", + "choice_a": "de", + "choice_b": "zh-CN", + "choice_c": "it", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1937 + }, + { + "path": "common_voice_en_19580621.mp3", + "question": "Recognize the language used by the speaker in this speech.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1938 + }, + { + "path": "common_voice_es_19139955.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "fr", + "choice_b": "es", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1939 + }, + { + "path": "common_voice_fr_19717928.mp3", + "question": "Listen to the speaker. Which language is it?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "fr", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1940 + }, + { + "path": "common_voice_es_19965527.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "en", + "choice_b": "it", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1941 + }, + { + "path": "common_voice_fr_19718881.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1942 + }, + { + "path": "common_voice_fr_20032634.mp3", + "question": "Identify the language in which the speech was told.", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1943 + }, + { + "path": "common_voice_es_19597168.mp3", + "question": "Listen to the speech. In which language is it told?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1944 + }, + { + "path": "common_voice_es_19641187.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1945 + }, + { + "path": "common_voice_en_19944327.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1946 + }, + { + "path": "common_voice_it_20018449.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "es", + "choice_b": "it", + "choice_c": "de", + "choice_d": "ja", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1947 + }, + { + "path": "common_voice_en_19759858.mp3", + "question": "Guess the language in the audio.", + "choice_a": "en", + "choice_b": "zh-CN", + "choice_c": "de", + "choice_d": "it", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1948 + }, + { + "path": "common_voice_es_18598469.mp3", + "question": "From the dialogue you heard, can you tell the language?", + "choice_a": "de", + "choice_b": "es", + "choice_c": "fr", + "choice_d": "ja", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1949 + }, + { + "path": "common_voice_fr_18053457.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "it", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1950 + }, + { + "path": "common_voice_en_20052685.mp3", + "question": "Guess the language in the audio.", + "choice_a": "zh-CN", + "choice_b": "ja", + "choice_c": "en", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1951 + }, + { + "path": "common_voice_en_19959178.mp3", + "question": "Identify the language of the speech.", + "choice_a": "fr", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1952 + }, + { + "path": "common_voice_en_629083.mp3", + "question": "Identify the language of the speech.", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1953 + }, + { + "path": "common_voice_es_19011259.mp3", + "question": "What language is the speaker performing in?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1954 + }, + { + "path": "common_voice_de_19740201.mp3", + "question": "What is the native language of the person speaking in this recording?", + "choice_a": "zh-CN", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1955 + }, + { + "path": "common_voice_de_19202165.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "de", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "fr", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1956 + }, + { + "path": "common_voice_fr_19028290.mp3", + "question": "Which language is being spoken by the speaker?", + "choice_a": "en", + "choice_b": "fr", + "choice_c": "it", + "choice_d": "ja", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1957 + }, + { + "path": "common_voice_it_20053055.mp3", + "question": "Identify the language of the speech you've just listened to.", + "choice_a": "it", + "choice_b": "ja", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1958 + }, + { + "path": "common_voice_en_19715985.mp3", + "question": "In which language is this speech being given?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1959 + }, + { + "path": "common_voice_de_19129697.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "zh-CN", + "choice_b": "de", + "choice_c": "ja", + "choice_d": "it", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1960 + }, + { + "path": "common_voice_en_18640103.mp3", + "question": "In which language is this speaker spoken?", + "choice_a": "en", + "choice_b": "es", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1961 + }, + { + "path": "common_voice_es_19760759.mp3", + "question": "identify the language of this speech.", + "choice_a": "ja", + "choice_b": "es", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1962 + }, + { + "path": "common_voice_zh-CN_18978001.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "de", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "zh-CN", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1963 + }, + { + "path": "common_voice_it_18812588.mp3", + "question": "What language is the speech in?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1964 + }, + { + "path": "common_voice_en_20048696.mp3", + "question": "In which language was the announcement made?", + "choice_a": "ja", + "choice_b": "fr", + "choice_c": "zh-CN", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1965 + }, + { + "path": "common_voice_fr_17920907.mp3", + "question": "What language is the speaker using in this audio clip?", + "choice_a": "en", + "choice_b": "ja", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1966 + }, + { + "path": "common_voice_fr_19741259.mp3", + "question": "What language is the speaker using?", + "choice_a": "de", + "choice_b": "en", + "choice_c": "fr", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1967 + }, + { + "path": "common_voice_de_19702745.mp3", + "question": "Which language is used in this speech?", + "choice_a": "ja", + "choice_b": "de", + "choice_c": "es", + "choice_d": "en", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1968 + }, + { + "path": "common_voice_en_20024177.mp3", + "question": "Which language does the speaker speak?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "ja", + "choice_d": "en", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1969 + }, + { + "path": "common_voice_en_584536.mp3", + "question": "The speech you just heard is in what language?", + "choice_a": "en", + "choice_b": "de", + "choice_c": "zh-CN", + "choice_d": "es", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1970 + }, + { + "path": "common_voice_it_20014826.mp3", + "question": "Which language is the speaker using in the audio?", + "choice_a": "es", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "it", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1971 + }, + { + "path": "common_voice_it_19969698.mp3", + "question": "Which language is used by the speaker in this recording?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "fr", + "choice_d": "en", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1972 + }, + { + "path": "common_voice_de_19675615.mp3", + "question": "Listen to the speech. What language is it?", + "choice_a": "fr", + "choice_b": "ja", + "choice_c": "it", + "choice_d": "de", + "answer_gt": "de", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1973 + }, + { + "path": "common_voice_en_19621541.mp3", + "question": "In which language is this speech being given?", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "es", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1974 + }, + { + "path": "common_voice_en_19551056.mp3", + "question": "Guess the language of the speech.", + "choice_a": "zh-CN", + "choice_b": "en", + "choice_c": "ja", + "choice_d": "de", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1975 + }, + { + "path": "common_voice_fr_19523068.mp3", + "question": "In which language is this speech being given?", + "choice_a": "zh-CN", + "choice_b": "it", + "choice_c": "de", + "choice_d": "fr", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1976 + }, + { + "path": "common_voice_en_19654378.mp3", + "question": "In which language is this speech being given?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "ja", + "answer_gt": "en", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1977 + }, + { + "path": "common_voice_es_19617597.mp3", + "question": "The speech you listened to is in which language?", + "choice_a": "it", + "choice_b": "es", + "choice_c": "en", + "choice_d": "de", + "answer_gt": "es", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1978 + }, + { + "path": "common_voice_fr_19350354.mp3", + "question": "The speaker is speaking in which language?", + "choice_a": "it", + "choice_b": "fr", + "choice_c": "es", + "choice_d": "zh-CN", + "answer_gt": "fr", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1979 + }, + { + "path": "common_voice_it_20032130.mp3", + "question": "Which of the following languages matches the audio?", + "choice_a": "it", + "choice_b": "zh-CN", + "choice_c": "es", + "choice_d": "de", + "answer_gt": "it", + "task_name": "Spoken_Language_Identification", + "dataset_name": "covost2", + "uniq_id": 1980 + }, + { + "path": "common_voice_en_31500131.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1981 + }, + { + "path": "common_voice_en_21722729.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1982 + }, + { + "path": "common_voice_en_18182977.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1983 + }, + { + "path": "common_voice_en_18857333.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1984 + }, + { + "path": "common_voice_en_25126315.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1985 + }, + { + "path": "common_voice_en_18109357.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1986 + }, + { + "path": "common_voice_en_18093355.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1987 + }, + { + "path": "common_voice_en_18593745.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1988 + }, + { + "path": "common_voice_en_19945492.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1989 + }, + { + "path": "common_voice_en_33409250.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1990 + }, + { + "path": "common_voice_en_22951224.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1991 + }, + { + "path": "common_voice_en_20291233.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1992 + }, + { + "path": "common_voice_en_553506.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1993 + }, + { + "path": "common_voice_en_31596100.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1994 + }, + { + "path": "common_voice_en_37221637.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1995 + }, + { + "path": "common_voice_en_23827068.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1996 + }, + { + "path": "common_voice_en_19015151.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1997 + }, + { + "path": "common_voice_en_19739328.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1998 + }, + { + "path": "common_voice_en_22435265.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 1999 + }, + { + "path": "common_voice_en_19825401.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2000 + }, + { + "path": "common_voice_en_31562353.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2001 + }, + { + "path": "common_voice_en_36030672.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2002 + }, + { + "path": "common_voice_en_31599503.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2003 + }, + { + "path": "common_voice_en_495880.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2004 + }, + { + "path": "common_voice_en_18844018.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2005 + }, + { + "path": "common_voice_en_20178860.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2006 + }, + { + "path": "common_voice_en_31680894.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2007 + }, + { + "path": "common_voice_en_22766766.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2008 + }, + { + "path": "common_voice_en_19641358.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2009 + }, + { + "path": "common_voice_en_21195578.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2010 + }, + { + "path": "common_voice_en_18948012.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2011 + }, + { + "path": "common_voice_en_660793.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2012 + }, + { + "path": "common_voice_en_19480238.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2013 + }, + { + "path": "common_voice_en_19212949.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2014 + }, + { + "path": "common_voice_en_17426846.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2015 + }, + { + "path": "common_voice_en_17895366.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2016 + }, + { + "path": "common_voice_en_26129440.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2017 + }, + { + "path": "common_voice_en_19939781.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2018 + }, + { + "path": "common_voice_en_19779226.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2019 + }, + { + "path": "common_voice_en_23756204.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2020 + }, + { + "path": "common_voice_en_27846449.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2021 + }, + { + "path": "common_voice_en_187061.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2022 + }, + { + "path": "common_voice_en_22469599.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2023 + }, + { + "path": "common_voice_en_19680459.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2024 + }, + { + "path": "common_voice_en_627541.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2025 + }, + { + "path": "common_voice_en_17669825.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2026 + }, + { + "path": "common_voice_en_22605409.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2027 + }, + { + "path": "common_voice_en_19489428.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2028 + }, + { + "path": "common_voice_en_17622129.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2029 + }, + { + "path": "common_voice_en_24230074.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2030 + }, + { + "path": "common_voice_en_18545942.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2031 + }, + { + "path": "common_voice_en_31584058.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2032 + }, + { + "path": "common_voice_en_19100981.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2033 + }, + { + "path": "common_voice_en_35260777.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2034 + }, + { + "path": "common_voice_en_20000747.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2035 + }, + { + "path": "common_voice_en_23993139.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2036 + }, + { + "path": "common_voice_en_18849782.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2037 + }, + { + "path": "common_voice_en_26950526.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2038 + }, + { + "path": "common_voice_en_36612799.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2039 + }, + { + "path": "common_voice_en_17272412.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2040 + }, + { + "path": "common_voice_en_20937832.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2041 + }, + { + "path": "common_voice_en_21261870.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2042 + }, + { + "path": "common_voice_en_18707937.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2043 + }, + { + "path": "common_voice_en_18352993.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2044 + }, + { + "path": "common_voice_en_18712692.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2045 + }, + { + "path": "common_voice_en_36520201.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2046 + }, + { + "path": "common_voice_en_1742051.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2047 + }, + { + "path": "common_voice_en_23739138.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2048 + }, + { + "path": "common_voice_en_22927465.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2049 + }, + { + "path": "common_voice_en_27070191.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2050 + }, + { + "path": "common_voice_en_17914760.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2051 + }, + { + "path": "common_voice_en_508297.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2052 + }, + { + "path": "common_voice_en_64837.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2053 + }, + { + "path": "common_voice_en_20036343.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2054 + }, + { + "path": "common_voice_en_17881653.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2055 + }, + { + "path": "common_voice_en_19653320.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2056 + }, + { + "path": "common_voice_en_31557573.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2057 + }, + { + "path": "common_voice_en_31611581.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2058 + }, + { + "path": "common_voice_en_18070351.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2059 + }, + { + "path": "common_voice_en_18489793.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2060 + }, + { + "path": "common_voice_en_34571828.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2061 + }, + { + "path": "common_voice_en_17357576.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2062 + }, + { + "path": "common_voice_en_31328735.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2063 + }, + { + "path": "common_voice_en_19532959.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2064 + }, + { + "path": "common_voice_en_19015148.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2065 + }, + { + "path": "common_voice_en_20022714.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2066 + }, + { + "path": "common_voice_en_19685105.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2067 + }, + { + "path": "common_voice_en_159722.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2068 + }, + { + "path": "common_voice_en_18342344.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2069 + }, + { + "path": "common_voice_en_31752092.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2070 + }, + { + "path": "common_voice_en_21092395.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2071 + }, + { + "path": "common_voice_en_31703154.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2072 + }, + { + "path": "common_voice_en_19169722.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2073 + }, + { + "path": "common_voice_en_33787625.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2074 + }, + { + "path": "common_voice_en_30535394.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2075 + }, + { + "path": "common_voice_en_29346877.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2076 + }, + { + "path": "common_voice_en_30424384.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2077 + }, + { + "path": "common_voice_en_31041412.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2078 + }, + { + "path": "common_voice_en_20930832.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2079 + }, + { + "path": "common_voice_en_20586173.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2080 + }, + { + "path": "common_voice_en_25884084.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2081 + }, + { + "path": "common_voice_en_34298351.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2082 + }, + { + "path": "common_voice_en_8805513.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2083 + }, + { + "path": "common_voice_en_19945306.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2084 + }, + { + "path": "common_voice_en_27734891.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2085 + }, + { + "path": "common_voice_en_22790803.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2086 + }, + { + "path": "common_voice_en_31755671.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2087 + }, + { + "path": "common_voice_en_31266588.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2088 + }, + { + "path": "common_voice_en_35026745.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2089 + }, + { + "path": "common_voice_en_32321230.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2090 + }, + { + "path": "common_voice_en_20261314.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2091 + }, + { + "path": "common_voice_en_17291039.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2092 + }, + { + "path": "common_voice_en_1323348.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2093 + }, + { + "path": "common_voice_en_20688736.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2094 + }, + { + "path": "common_voice_en_19311994.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2095 + }, + { + "path": "common_voice_en_17861959.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2096 + }, + { + "path": "common_voice_en_19935865.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2097 + }, + { + "path": "common_voice_en_27846447.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2098 + }, + { + "path": "common_voice_en_27619474.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2099 + }, + { + "path": "common_voice_en_32640268.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2100 + }, + { + "path": "common_voice_en_31578824.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2101 + }, + { + "path": "common_voice_en_34288692.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2102 + }, + { + "path": "common_voice_en_23896967.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2103 + }, + { + "path": "common_voice_en_19936931.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2104 + }, + { + "path": "common_voice_en_37040119.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2105 + }, + { + "path": "common_voice_en_17902010.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2106 + }, + { + "path": "common_voice_en_21667418.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2107 + }, + { + "path": "common_voice_en_23768958.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2108 + }, + { + "path": "common_voice_en_31206949.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2109 + }, + { + "path": "common_voice_en_17663996.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2110 + }, + { + "path": "common_voice_en_30580118.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2111 + }, + { + "path": "common_voice_en_18168851.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2112 + }, + { + "path": "common_voice_en_34917235.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2113 + }, + { + "path": "common_voice_en_18828401.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2114 + }, + { + "path": "common_voice_en_31572606.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2115 + }, + { + "path": "common_voice_en_659600.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2116 + }, + { + "path": "common_voice_en_31734340.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2117 + }, + { + "path": "common_voice_en_17285876.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2118 + }, + { + "path": "common_voice_en_21400996.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2119 + }, + { + "path": "common_voice_en_25021926.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2120 + }, + { + "path": "common_voice_en_17846037.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2121 + }, + { + "path": "common_voice_en_19546213.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2122 + }, + { + "path": "common_voice_en_26830417.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2123 + }, + { + "path": "common_voice_en_30406494.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2124 + }, + { + "path": "common_voice_en_32914884.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2125 + }, + { + "path": "common_voice_en_19954083.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2126 + }, + { + "path": "common_voice_en_25722369.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2127 + }, + { + "path": "common_voice_en_31754883.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2128 + }, + { + "path": "common_voice_en_19729823.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2129 + }, + { + "path": "common_voice_en_31684634.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2130 + }, + { + "path": "common_voice_en_18070362.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2131 + }, + { + "path": "common_voice_en_17259866.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2132 + }, + { + "path": "common_voice_en_18262068.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2133 + }, + { + "path": "common_voice_en_32859682.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2134 + }, + { + "path": "common_voice_en_19149077.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2135 + }, + { + "path": "common_voice_en_17938003.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2136 + }, + { + "path": "common_voice_en_18345208.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2137 + }, + { + "path": "common_voice_en_28847861.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2138 + }, + { + "path": "common_voice_en_18723995.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2139 + }, + { + "path": "common_voice_en_322482.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2140 + }, + { + "path": "common_voice_en_21858027.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2141 + }, + { + "path": "common_voice_en_21069613.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2142 + }, + { + "path": "common_voice_en_21679992.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2143 + }, + { + "path": "common_voice_en_18425397.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2144 + }, + { + "path": "common_voice_en_20258935.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2145 + }, + { + "path": "common_voice_en_32642819.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2146 + }, + { + "path": "common_voice_en_23646368.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2147 + }, + { + "path": "common_voice_en_20243166.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2148 + }, + { + "path": "common_voice_en_595873.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2149 + }, + { + "path": "common_voice_en_25898320.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2150 + }, + { + "path": "common_voice_en_567147.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2151 + }, + { + "path": "common_voice_en_18849783.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2152 + }, + { + "path": "common_voice_en_31609934.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2153 + }, + { + "path": "common_voice_en_31582710.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2154 + }, + { + "path": "common_voice_en_32929052.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2155 + }, + { + "path": "common_voice_en_19351145.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2156 + }, + { + "path": "common_voice_en_22802253.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2157 + }, + { + "path": "common_voice_en_19751836.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2158 + }, + { + "path": "common_voice_en_32092487.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2159 + }, + { + "path": "common_voice_en_31643995.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2160 + }, + { + "path": "common_voice_en_18813644.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2161 + }, + { + "path": "common_voice_en_552057.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2162 + }, + { + "path": "common_voice_en_20172863.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2163 + }, + { + "path": "common_voice_en_27257382.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2164 + }, + { + "path": "common_voice_en_331937.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2165 + }, + { + "path": "common_voice_en_20678933.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2166 + }, + { + "path": "common_voice_en_20987283.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2167 + }, + { + "path": "common_voice_en_27267612.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2168 + }, + { + "path": "common_voice_en_22653348.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2169 + }, + { + "path": "common_voice_en_31555005.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2170 + }, + { + "path": "common_voice_en_37118169.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2171 + }, + { + "path": "common_voice_en_491330.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2172 + }, + { + "path": "common_voice_en_31769648.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2173 + }, + { + "path": "common_voice_en_23655919.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2174 + }, + { + "path": "common_voice_en_672261.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2175 + }, + { + "path": "common_voice_en_19524811.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2176 + }, + { + "path": "common_voice_en_19774282.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2177 + }, + { + "path": "common_voice_en_26358.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2178 + }, + { + "path": "common_voice_en_675200.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2179 + }, + { + "path": "common_voice_en_16659359.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2180 + }, + { + "path": "common_voice_en_18948010.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2181 + }, + { + "path": "common_voice_en_18198686.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2182 + }, + { + "path": "common_voice_en_33606187.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2183 + }, + { + "path": "common_voice_en_21883920.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2184 + }, + { + "path": "common_voice_en_19742273.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2185 + }, + { + "path": "common_voice_en_19933879.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2186 + }, + { + "path": "common_voice_en_22717229.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2187 + }, + { + "path": "common_voice_en_37016367.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2188 + }, + { + "path": "common_voice_en_19545133.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2189 + }, + { + "path": "common_voice_en_19696556.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2190 + }, + { + "path": "common_voice_en_32280983.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2191 + }, + { + "path": "common_voice_en_17266327.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2192 + }, + { + "path": "common_voice_en_34891885.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2193 + }, + { + "path": "common_voice_en_27731438.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2194 + }, + { + "path": "common_voice_en_33550583.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2195 + }, + { + "path": "common_voice_en_31866896.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2196 + }, + { + "path": "common_voice_en_19956821.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2197 + }, + { + "path": "common_voice_en_573140.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2198 + }, + { + "path": "common_voice_en_31704085.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2199 + }, + { + "path": "common_voice_en_30951482.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2200 + }, + { + "path": "common_voice_en_19948136.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2201 + }, + { + "path": "common_voice_en_33433657.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2202 + }, + { + "path": "common_voice_en_23609781.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2203 + }, + { + "path": "common_voice_en_19041345.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2204 + }, + { + "path": "common_voice_en_36878640.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2205 + }, + { + "path": "common_voice_en_19222168.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2206 + }, + { + "path": "common_voice_en_21702346.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2207 + }, + { + "path": "common_voice_en_35772659.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2208 + }, + { + "path": "common_voice_en_17283107.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2209 + }, + { + "path": "common_voice_en_22951223.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2210 + }, + { + "path": "common_voice_en_17151232.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2211 + }, + { + "path": "common_voice_en_32171854.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2212 + }, + { + "path": "common_voice_en_19735705.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2213 + }, + { + "path": "common_voice_en_20182237.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2214 + }, + { + "path": "common_voice_en_32133713.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2215 + }, + { + "path": "common_voice_en_23852947.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2216 + }, + { + "path": "common_voice_en_498245.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2217 + }, + { + "path": "common_voice_en_22926410.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2218 + }, + { + "path": "common_voice_en_18273546.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2219 + }, + { + "path": "common_voice_en_25619619.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2220 + }, + { + "path": "common_voice_en_23739136.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2221 + }, + { + "path": "common_voice_en_13483785.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2222 + }, + { + "path": "common_voice_en_18979305.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2223 + }, + { + "path": "common_voice_en_36876338.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2224 + }, + { + "path": "common_voice_en_32311201.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2225 + }, + { + "path": "common_voice_en_19270939.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2226 + }, + { + "path": "common_voice_en_33371429.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2227 + }, + { + "path": "common_voice_en_32538389.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2228 + }, + { + "path": "common_voice_en_368765.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2229 + }, + { + "path": "common_voice_en_31661608.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2230 + }, + { + "path": "common_voice_en_373477.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2231 + }, + { + "path": "common_voice_en_34399333.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2232 + }, + { + "path": "common_voice_en_27386232.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2233 + }, + { + "path": "common_voice_en_32144886.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2234 + }, + { + "path": "common_voice_en_19988809.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2235 + }, + { + "path": "common_voice_en_22588894.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2236 + }, + { + "path": "common_voice_en_28197049.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2237 + }, + { + "path": "common_voice_en_20937834.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2238 + }, + { + "path": "common_voice_en_20478855.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2239 + }, + { + "path": "common_voice_en_20304833.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2240 + }, + { + "path": "common_voice_en_17938608.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2241 + }, + { + "path": "common_voice_en_27734771.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2242 + }, + { + "path": "common_voice_en_36538485.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2243 + }, + { + "path": "common_voice_en_19961336.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2244 + }, + { + "path": "common_voice_en_31711008.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2245 + }, + { + "path": "common_voice_en_27734770.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2246 + }, + { + "path": "common_voice_en_35714315.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2247 + }, + { + "path": "common_voice_en_19520255.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2248 + }, + { + "path": "common_voice_en_22709694.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2249 + }, + { + "path": "common_voice_en_20292104.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2250 + }, + { + "path": "common_voice_en_18661022.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2251 + }, + { + "path": "common_voice_en_17820439.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2252 + }, + { + "path": "common_voice_en_19962548.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2253 + }, + { + "path": "common_voice_en_17297426.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2254 + }, + { + "path": "common_voice_en_19715449.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2255 + }, + { + "path": "common_voice_en_17492946.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2256 + }, + { + "path": "common_voice_en_6225638.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2257 + }, + { + "path": "common_voice_en_123648.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2258 + }, + { + "path": "common_voice_en_5735881.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2259 + }, + { + "path": "common_voice_en_19031179.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2260 + }, + { + "path": "common_voice_en_20542865.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2261 + }, + { + "path": "common_voice_en_35092722.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2262 + }, + { + "path": "common_voice_en_32968320.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2263 + }, + { + "path": "common_voice_en_10564.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2264 + }, + { + "path": "common_voice_en_31582705.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2265 + }, + { + "path": "common_voice_en_30321989.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2266 + }, + { + "path": "common_voice_en_20907101.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2267 + }, + { + "path": "common_voice_en_18554229.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2268 + }, + { + "path": "common_voice_en_19966755.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2269 + }, + { + "path": "common_voice_en_27654278.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2270 + }, + { + "path": "common_voice_en_25618450.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2271 + }, + { + "path": "common_voice_en_33017669.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2272 + }, + { + "path": "common_voice_en_17891685.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2273 + }, + { + "path": "common_voice_en_19105943.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2274 + }, + { + "path": "common_voice_en_21265378.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2275 + }, + { + "path": "common_voice_en_27416438.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2276 + }, + { + "path": "common_voice_en_31695208.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2277 + }, + { + "path": "common_voice_en_31720201.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2278 + }, + { + "path": "common_voice_en_18967266.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2279 + }, + { + "path": "common_voice_en_35092634.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2280 + }, + { + "path": "common_voice_en_512011.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2281 + }, + { + "path": "common_voice_en_551377.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2282 + }, + { + "path": "common_voice_en_18317514.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2283 + }, + { + "path": "common_voice_en_17877327.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2284 + }, + { + "path": "common_voice_en_36398481.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2285 + }, + { + "path": "common_voice_en_32573474.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2286 + }, + { + "path": "common_voice_en_21349473.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2287 + }, + { + "path": "common_voice_en_463888.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2288 + }, + { + "path": "common_voice_en_19373318.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2289 + }, + { + "path": "common_voice_en_33389072.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2290 + }, + { + "path": "common_voice_en_17647993.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2291 + }, + { + "path": "common_voice_en_19939049.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2292 + }, + { + "path": "common_voice_en_18506356.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2293 + }, + { + "path": "common_voice_en_18166798.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2294 + }, + { + "path": "common_voice_en_35324674.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2295 + }, + { + "path": "common_voice_en_17761046.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2296 + }, + { + "path": "common_voice_en_30663927.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2297 + }, + { + "path": "common_voice_en_20296862.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2298 + }, + { + "path": "common_voice_en_18983327.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2299 + }, + { + "path": "common_voice_en_22730454.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2300 + }, + { + "path": "common_voice_en_595872.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2301 + }, + { + "path": "common_voice_en_20374635.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2302 + }, + { + "path": "common_voice_en_22325217.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2303 + }, + { + "path": "common_voice_en_553505.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2304 + }, + { + "path": "common_voice_en_19725350.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2305 + }, + { + "path": "common_voice_en_27102746.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2306 + }, + { + "path": "common_voice_en_37230615.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2307 + }, + { + "path": "common_voice_en_25716466.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2308 + }, + { + "path": "common_voice_en_27658680.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2309 + }, + { + "path": "common_voice_en_18309499.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2310 + }, + { + "path": "common_voice_en_629809.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2311 + }, + { + "path": "common_voice_en_20440160.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2312 + }, + { + "path": "common_voice_en_35437451.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2313 + }, + { + "path": "common_voice_en_31691583.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2314 + }, + { + "path": "common_voice_en_19693120.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2315 + }, + { + "path": "common_voice_en_17762228.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2316 + }, + { + "path": "common_voice_en_19936681.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2317 + }, + { + "path": "common_voice_en_18204116.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2318 + }, + { + "path": "common_voice_en_22925991.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2319 + }, + { + "path": "common_voice_en_19715450.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2320 + }, + { + "path": "common_voice_en_20178859.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2321 + }, + { + "path": "common_voice_en_37103360.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2322 + }, + { + "path": "common_voice_en_32189831.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2323 + }, + { + "path": "common_voice_en_18011378.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2324 + }, + { + "path": "common_voice_en_19703285.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2325 + }, + { + "path": "common_voice_en_34321814.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2326 + }, + { + "path": "common_voice_en_24437825.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2327 + }, + { + "path": "common_voice_en_18976653.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2328 + }, + { + "path": "common_voice_en_33195718.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2329 + }, + { + "path": "common_voice_en_19742276.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2330 + }, + { + "path": "common_voice_en_265829.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2331 + }, + { + "path": "common_voice_en_19643569.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2332 + }, + { + "path": "common_voice_en_31539933.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2333 + }, + { + "path": "common_voice_en_255533.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2334 + }, + { + "path": "common_voice_en_17354534.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2335 + }, + { + "path": "common_voice_en_18725234.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2336 + }, + { + "path": "common_voice_en_25705539.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2337 + }, + { + "path": "common_voice_en_31690078.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2338 + }, + { + "path": "common_voice_en_36538486.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2339 + }, + { + "path": "common_voice_en_28807933.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2340 + }, + { + "path": "common_voice_en_30513358.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2341 + }, + { + "path": "common_voice_en_20714771.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2342 + }, + { + "path": "common_voice_en_18828400.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2343 + }, + { + "path": "common_voice_en_25542151.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2344 + }, + { + "path": "common_voice_en_32101019.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2345 + }, + { + "path": "common_voice_en_20688626.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2346 + }, + { + "path": "common_voice_en_31752094.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2347 + }, + { + "path": "common_voice_en_36068063.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2348 + }, + { + "path": "common_voice_en_26287636.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2349 + }, + { + "path": "common_voice_en_35920080.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2350 + }, + { + "path": "common_voice_en_31794080.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2351 + }, + { + "path": "common_voice_en_30389805.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2352 + }, + { + "path": "common_voice_en_150709.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2353 + }, + { + "path": "common_voice_en_17271034.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2354 + }, + { + "path": "common_voice_en_20258933.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2355 + }, + { + "path": "common_voice_en_18099367.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2356 + }, + { + "path": "common_voice_en_19937494.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2357 + }, + { + "path": "common_voice_en_19692388.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2358 + }, + { + "path": "common_voice_en_19756110.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2359 + }, + { + "path": "common_voice_en_36938662.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2360 + }, + { + "path": "common_voice_en_36754042.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2361 + }, + { + "path": "common_voice_en_17956817.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2362 + }, + { + "path": "common_voice_en_23399565.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2363 + }, + { + "path": "common_voice_en_25031275.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2364 + }, + { + "path": "common_voice_en_584269.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2365 + }, + { + "path": "common_voice_en_21572998.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2366 + }, + { + "path": "common_voice_en_603726.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2367 + }, + { + "path": "common_voice_en_474450.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2368 + }, + { + "path": "common_voice_en_31798558.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2369 + }, + { + "path": "common_voice_en_16692475.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2370 + }, + { + "path": "common_voice_en_516191.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2371 + }, + { + "path": "common_voice_en_31572632.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2372 + }, + { + "path": "common_voice_en_17279812.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2373 + }, + { + "path": "common_voice_en_18332537.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2374 + }, + { + "path": "common_voice_en_19941596.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2375 + }, + { + "path": "common_voice_en_34693.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2376 + }, + { + "path": "common_voice_en_19396203.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2377 + }, + { + "path": "common_voice_en_26717743.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2378 + }, + { + "path": "common_voice_en_17857149.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2379 + }, + { + "path": "common_voice_en_36527772.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2380 + }, + { + "path": "common_voice_en_22967211.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2381 + }, + { + "path": "common_voice_en_19152834.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2382 + }, + { + "path": "common_voice_en_9684911.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2383 + }, + { + "path": "common_voice_en_17779717.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2384 + }, + { + "path": "common_voice_en_35148452.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2385 + }, + { + "path": "common_voice_en_24261272.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2386 + }, + { + "path": "common_voice_en_31534152.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2387 + }, + { + "path": "common_voice_en_37083044.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2388 + }, + { + "path": "common_voice_en_28826295.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2389 + }, + { + "path": "common_voice_en_20529061.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2390 + }, + { + "path": "common_voice_en_23805514.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2391 + }, + { + "path": "common_voice_en_27067974.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2392 + }, + { + "path": "common_voice_en_23379809.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2393 + }, + { + "path": "common_voice_en_177712.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2394 + }, + { + "path": "common_voice_en_20221794.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2395 + }, + { + "path": "common_voice_en_19609644.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2396 + }, + { + "path": "common_voice_en_16706008.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2397 + }, + { + "path": "common_voice_en_19729843.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2398 + }, + { + "path": "common_voice_en_23719019.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2399 + }, + { + "path": "common_voice_en_141797.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2400 + }, + { + "path": "common_voice_en_18848276.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2401 + }, + { + "path": "common_voice_en_20586574.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2402 + }, + { + "path": "common_voice_en_19547888.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2403 + }, + { + "path": "common_voice_en_23408793.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2404 + }, + { + "path": "common_voice_en_18278759.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2405 + }, + { + "path": "common_voice_en_691319.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2406 + }, + { + "path": "common_voice_en_86581.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2407 + }, + { + "path": "common_voice_en_18514959.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2408 + }, + { + "path": "common_voice_en_34842725.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2409 + }, + { + "path": "common_voice_en_22335218.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2410 + }, + { + "path": "common_voice_en_25337476.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2411 + }, + { + "path": "common_voice_en_573141.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2412 + }, + { + "path": "common_voice_en_31650544.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2413 + }, + { + "path": "common_voice_en_31692921.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2414 + }, + { + "path": "common_voice_en_31578412.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2415 + }, + { + "path": "common_voice_en_17152441.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2416 + }, + { + "path": "common_voice_en_250.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2417 + }, + { + "path": "common_voice_en_23995936.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2418 + }, + { + "path": "common_voice_en_19539287.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2419 + }, + { + "path": "common_voice_en_29440889.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2420 + }, + { + "path": "common_voice_en_36290075.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2421 + }, + { + "path": "common_voice_en_18758347.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2422 + }, + { + "path": "common_voice_en_36499457.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2423 + }, + { + "path": "common_voice_en_37016755.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2424 + }, + { + "path": "common_voice_en_17295580.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2425 + }, + { + "path": "common_voice_en_32592285.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2426 + }, + { + "path": "common_voice_en_31573009.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2427 + }, + { + "path": "common_voice_en_31788782.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2428 + }, + { + "path": "common_voice_en_106966.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2429 + }, + { + "path": "common_voice_en_32960342.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2430 + }, + { + "path": "common_voice_en_28831999.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2431 + }, + { + "path": "common_voice_en_28875056.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2432 + }, + { + "path": "common_voice_en_26912472.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2433 + }, + { + "path": "common_voice_en_20889724.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2434 + }, + { + "path": "common_voice_en_18124732.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2435 + }, + { + "path": "common_voice_en_17879862.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2436 + }, + { + "path": "common_voice_en_18276842.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2437 + }, + { + "path": "common_voice_en_19611935.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2438 + }, + { + "path": "common_voice_en_20282437.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2439 + }, + { + "path": "common_voice_en_560792.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2440 + }, + { + "path": "common_voice_en_21773347.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2441 + }, + { + "path": "common_voice_en_17249769.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2442 + }, + { + "path": "common_voice_en_21389622.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2443 + }, + { + "path": "common_voice_en_21092392.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2444 + }, + { + "path": "common_voice_en_20264170.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2445 + }, + { + "path": "common_voice_en_31754886.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2446 + }, + { + "path": "common_voice_en_26110325.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2447 + }, + { + "path": "common_voice_en_18355954.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2448 + }, + { + "path": "common_voice_en_27709191.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2449 + }, + { + "path": "common_voice_en_19182187.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2450 + }, + { + "path": "common_voice_en_31821767.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2451 + }, + { + "path": "common_voice_en_31612442.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2452 + }, + { + "path": "common_voice_en_3388.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2453 + }, + { + "path": "common_voice_en_20688689.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2454 + }, + { + "path": "common_voice_en_26523110.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2455 + }, + { + "path": "common_voice_en_18882314.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2456 + }, + { + "path": "common_voice_en_19396205.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2457 + }, + { + "path": "common_voice_en_1272597.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2458 + }, + { + "path": "common_voice_en_20002083.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2459 + }, + { + "path": "common_voice_en_34299867.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2460 + }, + { + "path": "common_voice_en_507708.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2461 + }, + { + "path": "common_voice_en_23768959.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2462 + }, + { + "path": "common_voice_en_17363890.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2463 + }, + { + "path": "common_voice_en_31571925.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2464 + }, + { + "path": "common_voice_en_25840867.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2465 + }, + { + "path": "common_voice_en_600880.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2466 + }, + { + "path": "common_voice_en_228006.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2467 + }, + { + "path": "common_voice_en_21367754.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2468 + }, + { + "path": "common_voice_en_17289324.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2469 + }, + { + "path": "common_voice_en_19945308.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2470 + }, + { + "path": "common_voice_en_17298604.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2471 + }, + { + "path": "common_voice_en_18579362.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2472 + }, + { + "path": "common_voice_en_522868.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2473 + }, + { + "path": "common_voice_en_18014311.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2474 + }, + { + "path": "common_voice_en_35526590.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2475 + }, + { + "path": "common_voice_en_25018118.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2476 + }, + { + "path": "common_voice_en_19940196.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2477 + }, + { + "path": "common_voice_en_19803152.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2478 + }, + { + "path": "common_voice_en_19937020.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2479 + }, + { + "path": "common_voice_en_609251.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2480 + }, + { + "path": "common_voice_en_31512826.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2481 + }, + { + "path": "common_voice_en_18339394.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2482 + }, + { + "path": "common_voice_en_20489055.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2483 + }, + { + "path": "common_voice_en_30741425.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2484 + }, + { + "path": "common_voice_en_19489423.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2485 + }, + { + "path": "common_voice_en_670563.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2486 + }, + { + "path": "common_voice_en_35400504.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2487 + }, + { + "path": "common_voice_en_18856231.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2488 + }, + { + "path": "common_voice_en_22334089.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2489 + }, + { + "path": "common_voice_en_19625477.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2490 + }, + { + "path": "common_voice_en_37236607.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2491 + }, + { + "path": "common_voice_en_27406749.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2492 + }, + { + "path": "common_voice_en_29984396.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2493 + }, + { + "path": "common_voice_en_18242955.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2494 + }, + { + "path": "common_voice_en_22727684.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2495 + }, + { + "path": "common_voice_en_21685241.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2496 + }, + { + "path": "common_voice_en_21403094.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2497 + }, + { + "path": "common_voice_en_22338655.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2498 + }, + { + "path": "common_voice_en_24278994.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2499 + }, + { + "path": "common_voice_en_30713099.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2500 + }, + { + "path": "common_voice_en_35007766.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2501 + }, + { + "path": "common_voice_en_19553273.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2502 + }, + { + "path": "common_voice_en_35832645.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2503 + }, + { + "path": "common_voice_en_17747097.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2504 + }, + { + "path": "common_voice_en_30699821.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2505 + }, + { + "path": "common_voice_en_35432462.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2506 + }, + { + "path": "common_voice_en_33000196.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2507 + }, + { + "path": "common_voice_en_31676946.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2508 + }, + { + "path": "common_voice_en_21374777.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2509 + }, + { + "path": "common_voice_en_22948690.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2510 + }, + { + "path": "common_voice_en_22709018.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2511 + }, + { + "path": "common_voice_en_16665997.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2512 + }, + { + "path": "common_voice_en_28001452.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2513 + }, + { + "path": "common_voice_en_18041824.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2514 + }, + { + "path": "common_voice_en_36455486.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2515 + }, + { + "path": "common_voice_en_19752016.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2516 + }, + { + "path": "common_voice_en_20489056.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2517 + }, + { + "path": "common_voice_en_36305987.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2518 + }, + { + "path": "common_voice_en_19998229.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2519 + }, + { + "path": "common_voice_en_18553336.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2520 + }, + { + "path": "common_voice_en_329504.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2521 + }, + { + "path": "common_voice_en_31723270.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2522 + }, + { + "path": "common_voice_en_19703287.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2523 + }, + { + "path": "common_voice_en_19464263.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2524 + }, + { + "path": "common_voice_en_28807929.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2525 + }, + { + "path": "common_voice_en_28832002.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2526 + }, + { + "path": "common_voice_en_18607767.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2527 + }, + { + "path": "common_voice_en_27731439.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2528 + }, + { + "path": "common_voice_en_18669343.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2529 + }, + { + "path": "common_voice_en_31665253.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2530 + }, + { + "path": "common_voice_en_20548238.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2531 + }, + { + "path": "common_voice_en_18255775.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2532 + }, + { + "path": "common_voice_en_23813868.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2533 + }, + { + "path": "common_voice_en_19596806.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2534 + }, + { + "path": "common_voice_en_18306537.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2535 + }, + { + "path": "common_voice_en_28595142.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2536 + }, + { + "path": "common_voice_en_37029932.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2537 + }, + { + "path": "common_voice_en_31003678.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2538 + }, + { + "path": "common_voice_en_71769.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2539 + }, + { + "path": "common_voice_en_17359854.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2540 + }, + { + "path": "common_voice_en_19152839.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2541 + }, + { + "path": "common_voice_en_30560272.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2542 + }, + { + "path": "common_voice_en_25261.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2543 + }, + { + "path": "common_voice_en_36003968.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2544 + }, + { + "path": "common_voice_en_18332463.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2545 + }, + { + "path": "common_voice_en_18969138.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2546 + }, + { + "path": "common_voice_en_19762166.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2547 + }, + { + "path": "common_voice_en_170041.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2548 + }, + { + "path": "common_voice_en_17882305.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2549 + }, + { + "path": "common_voice_en_19650893.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2550 + }, + { + "path": "common_voice_en_367926.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2551 + }, + { + "path": "common_voice_en_32351068.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2552 + }, + { + "path": "common_voice_en_20261609.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2553 + }, + { + "path": "common_voice_en_31676949.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2554 + }, + { + "path": "common_voice_en_19935427.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2555 + }, + { + "path": "common_voice_en_19596804.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2556 + }, + { + "path": "common_voice_en_17832074.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2557 + }, + { + "path": "common_voice_en_19739321.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2558 + }, + { + "path": "common_voice_en_21340521.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2559 + }, + { + "path": "common_voice_en_29326449.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2560 + }, + { + "path": "common_voice_en_17912298.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2561 + }, + { + "path": "common_voice_en_18813645.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2562 + }, + { + "path": "common_voice_en_24009284.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2563 + }, + { + "path": "common_voice_en_19693121.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2564 + }, + { + "path": "common_voice_en_19415544.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2565 + }, + { + "path": "common_voice_en_25646380.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2566 + }, + { + "path": "common_voice_en_19650921.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2567 + }, + { + "path": "common_voice_en_29123754.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2568 + }, + { + "path": "common_voice_en_19939156.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2569 + }, + { + "path": "common_voice_en_36383320.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2570 + }, + { + "path": "common_voice_en_18255018.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2571 + }, + { + "path": "common_voice_en_25635320.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2572 + }, + { + "path": "common_voice_en_20301324.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2573 + }, + { + "path": "common_voice_en_20304834.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2574 + }, + { + "path": "common_voice_en_20648476.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2575 + }, + { + "path": "common_voice_en_540648.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2576 + }, + { + "path": "common_voice_en_2731682.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2577 + }, + { + "path": "common_voice_en_18960599.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2578 + }, + { + "path": "common_voice_en_17250065.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2579 + }, + { + "path": "common_voice_en_603727.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2580 + }, + { + "path": "common_voice_en_22788366.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2581 + }, + { + "path": "common_voice_en_20278402.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2582 + }, + { + "path": "common_voice_en_17296984.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2583 + }, + { + "path": "common_voice_en_19266307.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2584 + }, + { + "path": "common_voice_en_30407406.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2585 + }, + { + "path": "common_voice_en_22713797.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2586 + }, + { + "path": "common_voice_en_37159307.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2587 + }, + { + "path": "common_voice_en_19647545.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2588 + }, + { + "path": "common_voice_en_31683731.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2589 + }, + { + "path": "common_voice_en_19729561.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2590 + }, + { + "path": "common_voice_en_560793.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2591 + }, + { + "path": "common_voice_en_20304587.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2592 + }, + { + "path": "common_voice_en_109017.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2593 + }, + { + "path": "common_voice_en_18506358.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2594 + }, + { + "path": "common_voice_en_23631965.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2595 + }, + { + "path": "common_voice_en_17281445.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2596 + }, + { + "path": "common_voice_en_623389.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2597 + }, + { + "path": "common_voice_en_32941670.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2598 + }, + { + "path": "common_voice_en_17285889.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2599 + }, + { + "path": "common_voice_en_18238070.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2600 + }, + { + "path": "common_voice_en_31705666.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2601 + }, + { + "path": "common_voice_en_21284877.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2602 + }, + { + "path": "common_voice_en_18218583.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2603 + }, + { + "path": "common_voice_en_37005755.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2604 + }, + { + "path": "common_voice_en_18733382.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2605 + }, + { + "path": "common_voice_en_23964610.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2606 + }, + { + "path": "common_voice_en_27257379.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2607 + }, + { + "path": "common_voice_en_24142683.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2608 + }, + { + "path": "common_voice_en_18979306.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2609 + }, + { + "path": "common_voice_en_19533333.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2610 + }, + { + "path": "common_voice_en_35061606.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2611 + }, + { + "path": "common_voice_en_19496793.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2612 + }, + { + "path": "common_voice_en_679076.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2613 + }, + { + "path": "common_voice_en_19212401.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2614 + }, + { + "path": "common_voice_en_29984394.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2615 + }, + { + "path": "common_voice_en_31612438.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2616 + }, + { + "path": "common_voice_en_19941683.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2617 + }, + { + "path": "common_voice_en_18854899.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2618 + }, + { + "path": "common_voice_en_31693302.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2619 + }, + { + "path": "common_voice_en_17789645.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2620 + }, + { + "path": "common_voice_en_18274245.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2621 + }, + { + "path": "common_voice_en_31667122.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2622 + }, + { + "path": "common_voice_en_20202676.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2623 + }, + { + "path": "common_voice_en_36079718.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2624 + }, + { + "path": "common_voice_en_35700499.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2625 + }, + { + "path": "common_voice_en_36398479.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2626 + }, + { + "path": "common_voice_en_33344944.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2627 + }, + { + "path": "common_voice_en_19209768.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2628 + }, + { + "path": "common_voice_en_20022539.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2629 + }, + { + "path": "common_voice_en_33606184.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2630 + }, + { + "path": "common_voice_en_19933238.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2631 + }, + { + "path": "common_voice_en_17782101.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2632 + }, + { + "path": "common_voice_en_348928.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2633 + }, + { + "path": "common_voice_en_30633200.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2634 + }, + { + "path": "common_voice_en_20202672.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2635 + }, + { + "path": "common_voice_en_27697441.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2636 + }, + { + "path": "common_voice_en_19842768.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2637 + }, + { + "path": "common_voice_en_19732218.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2638 + }, + { + "path": "common_voice_en_32311207.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2639 + }, + { + "path": "common_voice_en_32366593.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2640 + }, + { + "path": "common_voice_en_19686592.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2641 + }, + { + "path": "common_voice_en_18710315.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2642 + }, + { + "path": "common_voice_en_21615063.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2643 + }, + { + "path": "common_voice_en_21333801.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2644 + }, + { + "path": "common_voice_en_31512831.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2645 + }, + { + "path": "common_voice_en_17849979.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2646 + }, + { + "path": "common_voice_en_32801005.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2647 + }, + { + "path": "common_voice_en_21208382.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2648 + }, + { + "path": "common_voice_en_22866439.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2649 + }, + { + "path": "common_voice_en_18592575.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2650 + }, + { + "path": "common_voice_en_21798355.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2651 + }, + { + "path": "common_voice_en_33452651.mp3", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2652 + }, + { + "path": "common_voice_en_25187523.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2653 + }, + { + "path": "common_voice_en_22713796.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2654 + }, + { + "path": "common_voice_en_17257752.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2655 + }, + { + "path": "common_voice_en_18737952.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2656 + }, + { + "path": "common_voice_en_17294098.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2657 + }, + { + "path": "common_voice_en_17285890.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2658 + }, + { + "path": "common_voice_en_31571928.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2659 + }, + { + "path": "common_voice_en_31757413.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2660 + }, + { + "path": "common_voice_en_17251659.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2661 + }, + { + "path": "common_voice_en_19613456.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2662 + }, + { + "path": "common_voice_en_19941663.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2663 + }, + { + "path": "common_voice_en_17267531.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2664 + }, + { + "path": "common_voice_en_19652891.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2665 + }, + { + "path": "common_voice_en_35011971.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2666 + }, + { + "path": "common_voice_en_22936685.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2667 + }, + { + "path": "common_voice_en_35697775.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2668 + }, + { + "path": "common_voice_en_25664687.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2669 + }, + { + "path": "common_voice_en_1175989.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2670 + }, + { + "path": "common_voice_en_30585685.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2671 + }, + { + "path": "common_voice_en_37237056.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2672 + }, + { + "path": "common_voice_en_20262304.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2673 + }, + { + "path": "common_voice_en_31794078.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2674 + }, + { + "path": "common_voice_en_19819850.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2675 + }, + { + "path": "common_voice_en_25660678.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2676 + }, + { + "path": "common_voice_en_24472733.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2677 + }, + { + "path": "common_voice_en_18553335.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2678 + }, + { + "path": "common_voice_en_31328734.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2679 + }, + { + "path": "common_voice_en_24261278.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2680 + }, + { + "path": "common_voice_en_20586174.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2681 + }, + { + "path": "common_voice_en_20282120.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2682 + }, + { + "path": "common_voice_en_18974670.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2683 + }, + { + "path": "common_voice_en_17562717.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2684 + }, + { + "path": "common_voice_en_31536322.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2685 + }, + { + "path": "common_voice_en_27535749.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2686 + }, + { + "path": "common_voice_en_24205067.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2687 + }, + { + "path": "common_voice_en_32171092.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2688 + }, + { + "path": "common_voice_en_18009689.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2689 + }, + { + "path": "common_voice_en_31709932.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2690 + }, + { + "path": "common_voice_en_31599117.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2691 + }, + { + "path": "common_voice_en_19774280.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2692 + }, + { + "path": "common_voice_en_20301609.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2693 + }, + { + "path": "common_voice_en_19650760.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2694 + }, + { + "path": "common_voice_en_19031181.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2695 + }, + { + "path": "common_voice_en_37220997.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2696 + }, + { + "path": "common_voice_en_27314954.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2697 + }, + { + "path": "common_voice_en_17845943.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2698 + }, + { + "path": "common_voice_en_21307080.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2699 + }, + { + "path": "common_voice_en_17902879.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2700 + }, + { + "path": "common_voice_en_31772042.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2701 + }, + { + "path": "common_voice_en_18250601.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2702 + }, + { + "path": "common_voice_en_19964357.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2703 + }, + { + "path": "common_voice_en_64836.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2704 + }, + { + "path": "common_voice_en_16047346.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2705 + }, + { + "path": "common_voice_en_20076228.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2706 + }, + { + "path": "common_voice_en_18547306.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2707 + }, + { + "path": "common_voice_en_19654552.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2708 + }, + { + "path": "common_voice_en_31512428.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2709 + }, + { + "path": "common_voice_en_16659358.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2710 + }, + { + "path": "common_voice_en_367927.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2711 + }, + { + "path": "common_voice_en_18958067.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2712 + }, + { + "path": "common_voice_en_19854770.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2713 + }, + { + "path": "common_voice_en_21154507.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2714 + }, + { + "path": "common_voice_en_20289.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2715 + }, + { + "path": "common_voice_en_17281452.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2716 + }, + { + "path": "common_voice_en_19442319.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2717 + }, + { + "path": "common_voice_en_31789415.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2718 + }, + { + "path": "common_voice_en_18838805.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2719 + }, + { + "path": "common_voice_en_36564534.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2720 + }, + { + "path": "common_voice_en_20972098.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2721 + }, + { + "path": "common_voice_en_458607.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2722 + }, + { + "path": "common_voice_en_21943181.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2723 + }, + { + "path": "common_voice_en_19686599.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2724 + }, + { + "path": "common_voice_en_18122999.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2725 + }, + { + "path": "common_voice_en_21294317.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2726 + }, + { + "path": "common_voice_en_36440668.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2727 + }, + { + "path": "common_voice_en_18262069.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2728 + }, + { + "path": "common_voice_en_413120.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2729 + }, + { + "path": "common_voice_en_33471260.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2730 + }, + { + "path": "common_voice_en_21340519.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2731 + }, + { + "path": "common_voice_en_24985909.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2732 + }, + { + "path": "common_voice_en_19738882.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2733 + }, + { + "path": "common_voice_en_141796.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2734 + }, + { + "path": "common_voice_en_17268507.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2735 + }, + { + "path": "common_voice_en_20688728.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2736 + }, + { + "path": "common_voice_en_19846920.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2737 + }, + { + "path": "common_voice_en_18707935.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2738 + }, + { + "path": "common_voice_en_18256197.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2739 + }, + { + "path": "common_voice_en_19622847.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2740 + }, + { + "path": "common_voice_en_20380244.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2741 + }, + { + "path": "common_voice_en_21522619.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2742 + }, + { + "path": "common_voice_en_20848164.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2743 + }, + { + "path": "common_voice_en_31601173.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2744 + }, + { + "path": "common_voice_en_606948.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2745 + }, + { + "path": "common_voice_en_21786429.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2746 + }, + { + "path": "common_voice_en_32267025.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2747 + }, + { + "path": "common_voice_en_17756336.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2748 + }, + { + "path": "common_voice_en_37172186.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2749 + }, + { + "path": "common_voice_en_21134919.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2750 + }, + { + "path": "common_voice_en_26072506.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2751 + }, + { + "path": "common_voice_en_21162513.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2752 + }, + { + "path": "common_voice_en_23777207.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2753 + }, + { + "path": "common_voice_en_20685733.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2754 + }, + { + "path": "common_voice_en_31538690.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2755 + }, + { + "path": "common_voice_en_35526587.mp3", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2756 + }, + { + "path": "common_voice_en_20140424.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2757 + }, + { + "path": "common_voice_en_25152047.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2758 + }, + { + "path": "common_voice_en_20900715.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2759 + }, + { + "path": "common_voice_en_673730.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2760 + }, + { + "path": "common_voice_en_17556043.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2761 + }, + { + "path": "common_voice_en_20688737.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2762 + }, + { + "path": "common_voice_en_17782099.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2763 + }, + { + "path": "common_voice_en_20273153.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2764 + }, + { + "path": "common_voice_en_23964609.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2765 + }, + { + "path": "common_voice_en_19210963.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2766 + }, + { + "path": "common_voice_en_193485.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2767 + }, + { + "path": "common_voice_en_19480240.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2768 + }, + { + "path": "common_voice_en_17249432.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2769 + }, + { + "path": "common_voice_en_187603.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2770 + }, + { + "path": "common_voice_en_17357047.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2771 + }, + { + "path": "common_voice_en_17424862.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2772 + }, + { + "path": "common_voice_en_22951225.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2773 + }, + { + "path": "common_voice_en_20529053.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2774 + }, + { + "path": "common_voice_en_31562356.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2775 + }, + { + "path": "common_voice_en_19617112.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2776 + }, + { + "path": "common_voice_en_31649071.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2777 + }, + { + "path": "common_voice_en_690243.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2778 + }, + { + "path": "common_voice_en_36522655.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2779 + }, + { + "path": "common_voice_en_31673450.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2780 + }, + { + "path": "common_voice_en_18668688.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2781 + }, + { + "path": "common_voice_en_33443861.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2782 + }, + { + "path": "common_voice_en_37236620.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2783 + }, + { + "path": "common_voice_en_23846481.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2784 + }, + { + "path": "common_voice_en_19941502.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2785 + }, + { + "path": "common_voice_en_21730661.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2786 + }, + { + "path": "common_voice_en_18506357.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2787 + }, + { + "path": "common_voice_en_17258663.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2788 + }, + { + "path": "common_voice_en_18547302.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2789 + }, + { + "path": "common_voice_en_37228635.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2790 + }, + { + "path": "common_voice_en_35714312.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2791 + }, + { + "path": "common_voice_en_22575527.mp3", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2792 + }, + { + "path": "common_voice_en_21522616.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2793 + }, + { + "path": "common_voice_en_17760198.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2794 + }, + { + "path": "common_voice_en_19606474.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2795 + }, + { + "path": "common_voice_en_17371751.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2796 + }, + { + "path": "common_voice_en_18355955.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2797 + }, + { + "path": "common_voice_en_18933160.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2798 + }, + { + "path": "common_voice_en_24739116.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2799 + }, + { + "path": "common_voice_en_19149076.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2800 + }, + { + "path": "common_voice_en_136405.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2801 + }, + { + "path": "common_voice_en_18545940.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2802 + }, + { + "path": "common_voice_en_36951323.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2803 + }, + { + "path": "common_voice_en_20868630.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2804 + }, + { + "path": "common_voice_en_18137150.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2805 + }, + { + "path": "common_voice_en_21709043.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2806 + }, + { + "path": "common_voice_en_22699569.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2807 + }, + { + "path": "common_voice_en_19396201.mp3", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2808 + }, + { + "path": "common_voice_en_19553272.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2809 + }, + { + "path": "common_voice_en_17336130.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2810 + }, + { + "path": "common_voice_en_32941672.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2811 + }, + { + "path": "common_voice_en_20440161.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2812 + }, + { + "path": "common_voice_en_671188.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2813 + }, + { + "path": "common_voice_en_17354535.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2814 + }, + { + "path": "common_voice_en_31662914.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2815 + }, + { + "path": "common_voice_en_19768272.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2816 + }, + { + "path": "common_voice_en_71770.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2817 + }, + { + "path": "common_voice_en_34890002.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2818 + }, + { + "path": "common_voice_en_20854413.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2819 + }, + { + "path": "common_voice_en_32285275.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2820 + }, + { + "path": "common_voice_en_18730818.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2821 + }, + { + "path": "common_voice_en_22926411.mp3", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2822 + }, + { + "path": "common_voice_en_18758926.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2823 + }, + { + "path": "common_voice_en_20262803.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2824 + }, + { + "path": "common_voice_en_32377446.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2825 + }, + { + "path": "common_voice_en_31654622.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2826 + }, + { + "path": "common_voice_en_18800185.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2827 + }, + { + "path": "common_voice_en_30981010.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2828 + }, + { + "path": "common_voice_en_18712690.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2829 + }, + { + "path": "common_voice_en_31610538.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2830 + }, + { + "path": "common_voice_en_32297787.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2831 + }, + { + "path": "common_voice_en_276369.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2832 + }, + { + "path": "common_voice_en_19735675.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2833 + }, + { + "path": "common_voice_en_35697776.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2834 + }, + { + "path": "common_voice_en_20264169.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2835 + }, + { + "path": "common_voice_en_19757703.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2836 + }, + { + "path": "common_voice_en_693986.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2837 + }, + { + "path": "common_voice_en_19768275.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2838 + }, + { + "path": "common_voice_en_17780603.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2839 + }, + { + "path": "common_voice_en_21761960.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2840 + }, + { + "path": "common_voice_en_35011969.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2841 + }, + { + "path": "common_voice_en_19412457.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2842 + }, + { + "path": "common_voice_en_32875324.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2843 + }, + { + "path": "common_voice_en_17258255.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2844 + }, + { + "path": "common_voice_en_18864859.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2845 + }, + { + "path": "common_voice_en_30477719.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2846 + }, + { + "path": "common_voice_en_20526312.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2847 + }, + { + "path": "common_voice_en_18211915.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2848 + }, + { + "path": "common_voice_en_20526314.mp3", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2849 + }, + { + "path": "common_voice_en_19489421.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2850 + }, + { + "path": "common_voice_en_19939176.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2851 + }, + { + "path": "common_voice_en_18471299.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2852 + }, + { + "path": "common_voice_en_33290227.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2853 + }, + { + "path": "common_voice_en_31522572.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2854 + }, + { + "path": "common_voice_en_19547898.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2855 + }, + { + "path": "common_voice_en_522869.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2856 + }, + { + "path": "common_voice_en_18661019.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2857 + }, + { + "path": "common_voice_en_19621124.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2858 + }, + { + "path": "common_voice_en_18214584.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2859 + }, + { + "path": "common_voice_en_16030401.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2860 + }, + { + "path": "common_voice_en_30676708.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2861 + }, + { + "path": "common_voice_en_17904706.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2862 + }, + { + "path": "common_voice_en_30580116.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2863 + }, + { + "path": "common_voice_en_21283793.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2864 + }, + { + "path": "common_voice_en_17289296.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2865 + }, + { + "path": "common_voice_en_491058.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2866 + }, + { + "path": "common_voice_en_18849407.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2867 + }, + { + "path": "common_voice_en_18127728.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2868 + }, + { + "path": "common_voice_en_31700185.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2869 + }, + { + "path": "common_voice_en_20890435.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2870 + }, + { + "path": "common_voice_en_17255702.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2871 + }, + { + "path": "common_voice_en_21071526.mp3", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2872 + }, + { + "path": "common_voice_en_627540.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2873 + }, + { + "path": "common_voice_en_37266647.mp3", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2874 + }, + { + "path": "common_voice_en_19768273.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2875 + }, + { + "path": "common_voice_en_34897397.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2876 + }, + { + "path": "common_voice_en_17272411.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2877 + }, + { + "path": "common_voice_en_35582638.mp3", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2878 + }, + { + "path": "common_voice_en_17414289.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2879 + }, + { + "path": "common_voice_en_18303527.mp3", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2880 + }, + { + "path": "common_voice_en_24925871.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2881 + }, + { + "path": "common_voice_en_19064780.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2882 + }, + { + "path": "common_voice_en_19735704.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2883 + }, + { + "path": "common_voice_en_30615188.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2884 + }, + { + "path": "common_voice_en_20203640.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2885 + }, + { + "path": "common_voice_en_37133946.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2886 + }, + { + "path": "common_voice_en_207025.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2887 + }, + { + "path": "common_voice_en_19706318.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2888 + }, + { + "path": "common_voice_en_214716.mp3", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2889 + }, + { + "path": "common_voice_en_29440894.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2890 + }, + { + "path": "common_voice_en_31433767.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2891 + }, + { + "path": "common_voice_en_31877034.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2892 + }, + { + "path": "common_voice_en_20688746.mp3", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2893 + }, + { + "path": "common_voice_en_22469604.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2894 + }, + { + "path": "common_voice_en_19288096.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2895 + }, + { + "path": "common_voice_en_17260337.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2896 + }, + { + "path": "common_voice_en_20688632.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2897 + }, + { + "path": "common_voice_en_24230076.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2898 + }, + { + "path": "common_voice_en_20283482.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2899 + }, + { + "path": "common_voice_en_35854111.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2900 + }, + { + "path": "common_voice_en_514622.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2901 + }, + { + "path": "common_voice_en_25650381.mp3", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2902 + }, + { + "path": "common_voice_en_19650923.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2903 + }, + { + "path": "common_voice_en_37083045.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2904 + }, + { + "path": "common_voice_en_32190743.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2905 + }, + { + "path": "common_voice_en_19270941.mp3", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2906 + }, + { + "path": "common_voice_en_17871648.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2907 + }, + { + "path": "common_voice_en_25423290.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2908 + }, + { + "path": "common_voice_en_211447.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2909 + }, + { + "path": "common_voice_en_21685243.mp3", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2910 + }, + { + "path": "common_voice_en_22883988.mp3", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2911 + }, + { + "path": "common_voice_en_23379807.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2912 + }, + { + "path": "common_voice_en_23749403.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2913 + }, + { + "path": "common_voice_en_17912300.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2914 + }, + { + "path": "common_voice_en_22740552.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2915 + }, + { + "path": "common_voice_en_459460.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2916 + }, + { + "path": "common_voice_en_17699273.mp3", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2917 + }, + { + "path": "common_voice_en_35798064.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2918 + }, + { + "path": "common_voice_en_21307079.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2919 + }, + { + "path": "common_voice_en_19794969.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2920 + }, + { + "path": "common_voice_en_17624849.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2921 + }, + { + "path": "common_voice_en_26357.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2922 + }, + { + "path": "common_voice_en_495881.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2923 + }, + { + "path": "common_voice_en_31865132.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2924 + }, + { + "path": "common_voice_en_21300205.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2925 + }, + { + "path": "common_voice_en_35376669.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2926 + }, + { + "path": "common_voice_en_522009.mp3", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2927 + }, + { + "path": "common_voice_en_1156671.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2928 + }, + { + "path": "common_voice_en_203202.mp3", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2929 + }, + { + "path": "common_voice_en_20466244.mp3", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2930 + }, + { + "path": "common_voice_en_31789465.mp3", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2931 + }, + { + "path": "common_voice_en_21264428.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2932 + }, + { + "path": "common_voice_en_32171863.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2933 + }, + { + "path": "common_voice_en_24142666.mp3", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2934 + }, + { + "path": "common_voice_en_19953165.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2935 + }, + { + "path": "common_voice_en_33280385.mp3", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2936 + }, + { + "path": "common_voice_en_19451125.mp3", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2937 + }, + { + "path": "common_voice_en_64648.mp3", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2938 + }, + { + "path": "common_voice_en_31662924.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2939 + }, + { + "path": "common_voice_en_30321993.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2940 + }, + { + "path": "common_voice_en_17880199.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2941 + }, + { + "path": "common_voice_en_18170949.mp3", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2942 + }, + { + "path": "common_voice_en_19624957.mp3", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2943 + }, + { + "path": "common_voice_en_18343224.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2944 + }, + { + "path": "common_voice_en_17934372.mp3", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2945 + }, + { + "path": "common_voice_en_18485414.mp3", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2946 + }, + { + "path": "common_voice_en_37280774.mp3", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2947 + }, + { + "path": "common_voice_en_19666439.mp3", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2948 + }, + { + "path": "common_voice_en_596851.mp3", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2949 + }, + { + "path": "common_voice_en_20221791.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2950 + }, + { + "path": "common_voice_en_32412791.mp3", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2951 + }, + { + "path": "common_voice_en_19700531.mp3", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2952 + }, + { + "path": "common_voice_en_19708705.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2953 + }, + { + "path": "common_voice_en_19725499.mp3", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2954 + }, + { + "path": "common_voice_en_26233296.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2955 + }, + { + "path": "common_voice_en_356183.mp3", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2956 + }, + { + "path": "common_voice_en_36003972.mp3", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2957 + }, + { + "path": "common_voice_en_31699299.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2958 + }, + { + "path": "common_voice_en_19011342.mp3", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2959 + }, + { + "path": "common_voice_en_37069470.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2960 + }, + { + "path": "common_voice_en_25941300.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2961 + }, + { + "path": "common_voice_en_31636383.mp3", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2962 + }, + { + "path": "common_voice_en_19532961.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2963 + }, + { + "path": "common_voice_en_19011343.mp3", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2964 + }, + { + "path": "common_voice_en_19981532.mp3", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2965 + }, + { + "path": "common_voice_en_26324665.mp3", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2966 + }, + { + "path": "common_voice_en_22588810.mp3", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2967 + }, + { + "path": "common_voice_en_18257271.mp3", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2968 + }, + { + "path": "common_voice_en_18295850.mp3", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2969 + }, + { + "path": "common_voice_en_33321330.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2970 + }, + { + "path": "common_voice_en_18727227.mp3", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2971 + }, + { + "path": "common_voice_en_17285501.mp3", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "male", + "choice_b": "female", + "answer_gt": "male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "common_voice_13_en", + "uniq_id": 2972 + }, + { + "path": "dia210_utt0.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2973 + }, + { + "path": "dia262_utt10.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2974 + }, + { + "path": "dia127_utt5.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2975 + }, + { + "path": "dia14_utt9.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2976 + }, + { + "path": "dia97_utt10.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2977 + }, + { + "path": "dia189_utt0.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2978 + }, + { + "path": "dia167_utt7.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2979 + }, + { + "path": "dia137_utt9.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2980 + }, + { + "path": "dia229_utt4.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2981 + }, + { + "path": "dia241_utt11.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2982 + }, + { + "path": "dia139_utt5.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2983 + }, + { + "path": "dia217_utt2.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2984 + }, + { + "path": "dia98_utt8.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2985 + }, + { + "path": "dia264_utt1.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2986 + }, + { + "path": "dia268_utt5.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2987 + }, + { + "path": "dia157_utt2.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2988 + }, + { + "path": "dia65_utt10.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2989 + }, + { + "path": "dia112_utt2.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2990 + }, + { + "path": "dia64_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2991 + }, + { + "path": "dia155_utt4.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2992 + }, + { + "path": "dia208_utt2.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2993 + }, + { + "path": "dia151_utt3.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2994 + }, + { + "path": "dia244_utt10.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2995 + }, + { + "path": "dia41_utt4.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2996 + }, + { + "path": "dia52_utt1.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2997 + }, + { + "path": "dia102_utt7.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2998 + }, + { + "path": "dia248_utt0.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 2999 + }, + { + "path": "dia196_utt0.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3000 + }, + { + "path": "dia105_utt0.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3001 + }, + { + "path": "dia156_utt3.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3002 + }, + { + "path": "dia229_utt1.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3003 + }, + { + "path": "dia206_utt0.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3004 + }, + { + "path": "dia173_utt15.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3005 + }, + { + "path": "dia155_utt2.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3006 + }, + { + "path": "dia150_utt1.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3007 + }, + { + "path": "dia25_utt4.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3008 + }, + { + "path": "dia139_utt7.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3009 + }, + { + "path": "dia261_utt4.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3010 + }, + { + "path": "dia237_utt5.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3011 + }, + { + "path": "dia75_utt2.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3012 + }, + { + "path": "dia253_utt10.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3013 + }, + { + "path": "dia69_utt3.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3014 + }, + { + "path": "dia168_utt1.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3015 + }, + { + "path": "dia222_utt8.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3016 + }, + { + "path": "dia252_utt1.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3017 + }, + { + "path": "dia278_utt3.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3018 + }, + { + "path": "dia253_utt12.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3019 + }, + { + "path": "dia7_utt7.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3020 + }, + { + "path": "dia191_utt0.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3021 + }, + { + "path": "dia201_utt4.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3022 + }, + { + "path": "dia29_utt12.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3023 + }, + { + "path": "dia100_utt14.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3024 + }, + { + "path": "dia168_utt13.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3025 + }, + { + "path": "dia138_utt2.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3026 + }, + { + "path": "dia143_utt4.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3027 + }, + { + "path": "dia119_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3028 + }, + { + "path": "dia136_utt0.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3029 + }, + { + "path": "dia185_utt5.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3030 + }, + { + "path": "dia85_utt3.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3031 + }, + { + "path": "dia56_utt1.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3032 + }, + { + "path": "dia73_utt2.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3033 + }, + { + "path": "dia166_utt11.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3034 + }, + { + "path": "dia167_utt5.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3035 + }, + { + "path": "dia242_utt3.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3036 + }, + { + "path": "dia247_utt3.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3037 + }, + { + "path": "dia130_utt17.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3038 + }, + { + "path": "dia7_utt8.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3039 + }, + { + "path": "dia277_utt0.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3040 + }, + { + "path": "dia132_utt16.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3041 + }, + { + "path": "dia107_utt1.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3042 + }, + { + "path": "dia71_utt9.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3043 + }, + { + "path": "dia154_utt5.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3044 + }, + { + "path": "dia186_utt1.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3045 + }, + { + "path": "dia279_utt7.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3046 + }, + { + "path": "dia244_utt12.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3047 + }, + { + "path": "dia260_utt4.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3048 + }, + { + "path": "dia271_utt5.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3049 + }, + { + "path": "dia163_utt4.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3050 + }, + { + "path": "dia132_utt19.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3051 + }, + { + "path": "dia71_utt11.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3052 + }, + { + "path": "dia223_utt5.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3053 + }, + { + "path": "dia109_utt6.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3054 + }, + { + "path": "dia100_utt0.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3055 + }, + { + "path": "dia123_utt1.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3056 + }, + { + "path": "dia96_utt16.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3057 + }, + { + "path": "dia196_utt17.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3058 + }, + { + "path": "dia160_utt12.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3059 + }, + { + "path": "dia110_utt7.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3060 + }, + { + "path": "dia137_utt3.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3061 + }, + { + "path": "dia77_utt2.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3062 + }, + { + "path": "dia166_utt6.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3063 + }, + { + "path": "dia196_utt13.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3064 + }, + { + "path": "dia125_utt21.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3065 + }, + { + "path": "dia65_utt8.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3066 + }, + { + "path": "dia76_utt12.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3067 + }, + { + "path": "dia233_utt7.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3068 + }, + { + "path": "dia108_utt6.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3069 + }, + { + "path": "dia17_utt10.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3070 + }, + { + "path": "dia195_utt2.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3071 + }, + { + "path": "dia173_utt4.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3072 + }, + { + "path": "dia52_utt11.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3073 + }, + { + "path": "dia270_utt6.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3074 + }, + { + "path": "dia68_utt6.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3075 + }, + { + "path": "dia78_utt0.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3076 + }, + { + "path": "dia195_utt4.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3077 + }, + { + "path": "dia137_utt2.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3078 + }, + { + "path": "dia120_utt12.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3079 + }, + { + "path": "dia261_utt7.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3080 + }, + { + "path": "dia11_utt3.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3081 + }, + { + "path": "dia30_utt4.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3082 + }, + { + "path": "dia76_utt11.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3083 + }, + { + "path": "dia49_utt5.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3084 + }, + { + "path": "dia222_utt1.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3085 + }, + { + "path": "dia164_utt6.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3086 + }, + { + "path": "dia204_utt2.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3087 + }, + { + "path": "dia184_utt5.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3088 + }, + { + "path": "dia175_utt18.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3089 + }, + { + "path": "dia184_utt3.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3090 + }, + { + "path": "dia240_utt9.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3091 + }, + { + "path": "dia198_utt9.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3092 + }, + { + "path": "dia192_utt0.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3093 + }, + { + "path": "dia99_utt2.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3094 + }, + { + "path": "dia20_utt2.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3095 + }, + { + "path": "dia210_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3096 + }, + { + "path": "dia57_utt0.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3097 + }, + { + "path": "dia169_utt4.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3098 + }, + { + "path": "dia77_utt4.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3099 + }, + { + "path": "dia199_utt7.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3100 + }, + { + "path": "dia144_utt1.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3101 + }, + { + "path": "dia279_utt4.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3102 + }, + { + "path": "dia120_utt13.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3103 + }, + { + "path": "dia208_utt13.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3104 + }, + { + "path": "dia20_utt6.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3105 + }, + { + "path": "dia26_utt0.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3106 + }, + { + "path": "dia201_utt10.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3107 + }, + { + "path": "dia111_utt5.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3108 + }, + { + "path": "dia56_utt3.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3109 + }, + { + "path": "dia147_utt5.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3110 + }, + { + "path": "dia266_utt3.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3111 + }, + { + "path": "dia76_utt17.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3112 + }, + { + "path": "dia41_utt12.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3113 + }, + { + "path": "dia14_utt6.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3114 + }, + { + "path": "dia49_utt14.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3115 + }, + { + "path": "dia125_utt8.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3116 + }, + { + "path": "dia231_utt13.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3117 + }, + { + "path": "dia9_utt13.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3118 + }, + { + "path": "dia95_utt13.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3119 + }, + { + "path": "dia221_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3120 + }, + { + "path": "dia20_utt7.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3121 + }, + { + "path": "dia121_utt4.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3122 + }, + { + "path": "dia266_utt11.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3123 + }, + { + "path": "dia278_utt8.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3124 + }, + { + "path": "dia59_utt0.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3125 + }, + { + "path": "dia169_utt5.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3126 + }, + { + "path": "dia140_utt14.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3127 + }, + { + "path": "dia48_utt6.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3128 + }, + { + "path": "dia77_utt1.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3129 + }, + { + "path": "dia212_utt0.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3130 + }, + { + "path": "dia140_utt5.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3131 + }, + { + "path": "dia32_utt8.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3132 + }, + { + "path": "dia199_utt12.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3133 + }, + { + "path": "dia109_utt1.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3134 + }, + { + "path": "dia210_utt4.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3135 + }, + { + "path": "dia97_utt6.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3136 + }, + { + "path": "dia120_utt11.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3137 + }, + { + "path": "dia56_utt11.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3138 + }, + { + "path": "dia153_utt4.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3139 + }, + { + "path": "dia119_utt1.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3140 + }, + { + "path": "dia268_utt10.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3141 + }, + { + "path": "dia9_utt3.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3142 + }, + { + "path": "dia86_utt3.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3143 + }, + { + "path": "dia100_utt19.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3144 + }, + { + "path": "dia163_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3145 + }, + { + "path": "dia237_utt4.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3146 + }, + { + "path": "dia254_utt11.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3147 + }, + { + "path": "dia102_utt3.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3148 + }, + { + "path": "dia278_utt6.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3149 + }, + { + "path": "dia49_utt4.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3150 + }, + { + "path": "dia157_utt4.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3151 + }, + { + "path": "dia210_utt6.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3152 + }, + { + "path": "dia175_utt4.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3153 + }, + { + "path": "dia45_utt3.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3154 + }, + { + "path": "dia88_utt1.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3155 + }, + { + "path": "dia110_utt3.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3156 + }, + { + "path": "dia57_utt4.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3157 + }, + { + "path": "dia16_utt0.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3158 + }, + { + "path": "dia279_utt9.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3159 + }, + { + "path": "dia250_utt7.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3160 + }, + { + "path": "dia267_utt10.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3161 + }, + { + "path": "dia152_utt2.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3162 + }, + { + "path": "dia48_utt0.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3163 + }, + { + "path": "dia146_utt6.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3164 + }, + { + "path": "dia254_utt13.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3165 + }, + { + "path": "dia119_utt6.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3166 + }, + { + "path": "dia199_utt10.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3167 + }, + { + "path": "dia279_utt3.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3168 + }, + { + "path": "dia193_utt1.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3169 + }, + { + "path": "dia96_utt17.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3170 + }, + { + "path": "dia232_utt8.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3171 + }, + { + "path": "dia70_utt11.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3172 + }, + { + "path": "dia98_utt7.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3173 + }, + { + "path": "dia71_utt5.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3174 + }, + { + "path": "dia102_utt6.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3175 + }, + { + "path": "dia125_utt10.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3176 + }, + { + "path": "dia157_utt0.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3177 + }, + { + "path": "dia6_utt1.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3178 + }, + { + "path": "dia125_utt19.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3179 + }, + { + "path": "dia239_utt1.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3180 + }, + { + "path": "dia113_utt10.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3181 + }, + { + "path": "dia132_utt14.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3182 + }, + { + "path": "dia217_utt3.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3183 + }, + { + "path": "dia15_utt3.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3184 + }, + { + "path": "dia257_utt1.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3185 + }, + { + "path": "dia64_utt6.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3186 + }, + { + "path": "dia88_utt2.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3187 + }, + { + "path": "dia56_utt0.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3188 + }, + { + "path": "dia64_utt11.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3189 + }, + { + "path": "dia193_utt4.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3190 + }, + { + "path": "dia232_utt5.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3191 + }, + { + "path": "dia169_utt11.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3192 + }, + { + "path": "dia111_utt14.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3193 + }, + { + "path": "dia188_utt3.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3194 + }, + { + "path": "dia146_utt0.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3195 + }, + { + "path": "dia215_utt0.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3196 + }, + { + "path": "dia203_utt5.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3197 + }, + { + "path": "dia32_utt6.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3198 + }, + { + "path": "dia126_utt14.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3199 + }, + { + "path": "dia149_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3200 + }, + { + "path": "dia83_utt10.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3201 + }, + { + "path": "dia232_utt4.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3202 + }, + { + "path": "dia29_utt4.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3203 + }, + { + "path": "dia199_utt2.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3204 + }, + { + "path": "dia125_utt17.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3205 + }, + { + "path": "dia92_utt5.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3206 + }, + { + "path": "dia208_utt6.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3207 + }, + { + "path": "dia217_utt11.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3208 + }, + { + "path": "dia163_utt0.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3209 + }, + { + "path": "dia220_utt1.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3210 + }, + { + "path": "dia22_utt18.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3211 + }, + { + "path": "dia226_utt0.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3212 + }, + { + "path": "dia252_utt3.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3213 + }, + { + "path": "dia229_utt6.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3214 + }, + { + "path": "dia83_utt2.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3215 + }, + { + "path": "dia168_utt11.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3216 + }, + { + "path": "dia165_utt0.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3217 + }, + { + "path": "dia202_utt6.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3218 + }, + { + "path": "dia201_utt2.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3219 + }, + { + "path": "dia174_utt12.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3220 + }, + { + "path": "dia86_utt7.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3221 + }, + { + "path": "dia3_utt1.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3222 + }, + { + "path": "dia139_utt6.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3223 + }, + { + "path": "dia103_utt0.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3224 + }, + { + "path": "dia160_utt8.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3225 + }, + { + "path": "dia192_utt3.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3226 + }, + { + "path": "dia279_utt14.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3227 + }, + { + "path": "dia68_utt12.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3228 + }, + { + "path": "dia278_utt1.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3229 + }, + { + "path": "dia104_utt0.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3230 + }, + { + "path": "dia47_utt1.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3231 + }, + { + "path": "dia215_utt4.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3232 + }, + { + "path": "dia85_utt2.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3233 + }, + { + "path": "dia161_utt10.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3234 + }, + { + "path": "dia255_utt4.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3235 + }, + { + "path": "dia17_utt27.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3236 + }, + { + "path": "dia185_utt1.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3237 + }, + { + "path": "dia51_utt0.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3238 + }, + { + "path": "dia22_utt7.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3239 + }, + { + "path": "dia241_utt5.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3240 + }, + { + "path": "dia76_utt13.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3241 + }, + { + "path": "dia96_utt0.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3242 + }, + { + "path": "dia173_utt5.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3243 + }, + { + "path": "dia5_utt0.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3244 + }, + { + "path": "dia270_utt0.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3245 + }, + { + "path": "dia29_utt11.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3246 + }, + { + "path": "dia10_utt1.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3247 + }, + { + "path": "dia96_utt8.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3248 + }, + { + "path": "dia263_utt4.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3249 + }, + { + "path": "dia97_utt8.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3250 + }, + { + "path": "dia14_utt0.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3251 + }, + { + "path": "dia45_utt6.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3252 + }, + { + "path": "dia205_utt6.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3253 + }, + { + "path": "dia78_utt9.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3254 + }, + { + "path": "dia112_utt21.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3255 + }, + { + "path": "dia242_utt2.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3256 + }, + { + "path": "dia174_utt15.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3257 + }, + { + "path": "dia116_utt10.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3258 + }, + { + "path": "dia180_utt4.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3259 + }, + { + "path": "dia53_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3260 + }, + { + "path": "dia1_utt0.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3261 + }, + { + "path": "dia228_utt9.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3262 + }, + { + "path": "dia140_utt3.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3263 + }, + { + "path": "dia205_utt1.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3264 + }, + { + "path": "dia137_utt13.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3265 + }, + { + "path": "dia203_utt9.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3266 + }, + { + "path": "dia33_utt5.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3267 + }, + { + "path": "dia110_utt1.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3268 + }, + { + "path": "dia244_utt15.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3269 + }, + { + "path": "dia99_utt5.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3270 + }, + { + "path": "dia124_utt7.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3271 + }, + { + "path": "dia65_utt6.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3272 + }, + { + "path": "dia160_utt2.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3273 + }, + { + "path": "dia191_utt6.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3274 + }, + { + "path": "dia73_utt6.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3275 + }, + { + "path": "dia211_utt3.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3276 + }, + { + "path": "dia184_utt12.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3277 + }, + { + "path": "dia225_utt4.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3278 + }, + { + "path": "dia124_utt5.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3279 + }, + { + "path": "dia229_utt3.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3280 + }, + { + "path": "dia107_utt12.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3281 + }, + { + "path": "dia245_utt2.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3282 + }, + { + "path": "dia48_utt2.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3283 + }, + { + "path": "dia53_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3284 + }, + { + "path": "dia25_utt16.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3285 + }, + { + "path": "dia261_utt1.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3286 + }, + { + "path": "dia255_utt7.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3287 + }, + { + "path": "dia76_utt7.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3288 + }, + { + "path": "dia176_utt1.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3289 + }, + { + "path": "dia139_utt9.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3290 + }, + { + "path": "dia271_utt4.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3291 + }, + { + "path": "dia43_utt5.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3292 + }, + { + "path": "dia245_utt0.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3293 + }, + { + "path": "dia60_utt6.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3294 + }, + { + "path": "dia22_utt12.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3295 + }, + { + "path": "dia240_utt11.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3296 + }, + { + "path": "dia42_utt3.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3297 + }, + { + "path": "dia252_utt7.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3298 + }, + { + "path": "dia100_utt1.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3299 + }, + { + "path": "dia108_utt5.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3300 + }, + { + "path": "dia131_utt1.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3301 + }, + { + "path": "dia22_utt21.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3302 + }, + { + "path": "dia208_utt12.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3303 + }, + { + "path": "dia15_utt5.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3304 + }, + { + "path": "dia12_utt3.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3305 + }, + { + "path": "dia47_utt11.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3306 + }, + { + "path": "dia264_utt0.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3307 + }, + { + "path": "dia74_utt3.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3308 + }, + { + "path": "dia125_utt0.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3309 + }, + { + "path": "dia153_utt0.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3310 + }, + { + "path": "dia231_utt7.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3311 + }, + { + "path": "dia147_utt2.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3312 + }, + { + "path": "dia253_utt5.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3313 + }, + { + "path": "dia133_utt15.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3314 + }, + { + "path": "dia146_utt1.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3315 + }, + { + "path": "dia171_utt3.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3316 + }, + { + "path": "dia99_utt6.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3317 + }, + { + "path": "dia237_utt6.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3318 + }, + { + "path": "dia42_utt1.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3319 + }, + { + "path": "dia90_utt1.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3320 + }, + { + "path": "dia52_utt0.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3321 + }, + { + "path": "dia237_utt3.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3322 + }, + { + "path": "dia155_utt0.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3323 + }, + { + "path": "dia100_utt2.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3324 + }, + { + "path": "dia52_utt2.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3325 + }, + { + "path": "dia190_utt4.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3326 + }, + { + "path": "dia207_utt0.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3327 + }, + { + "path": "dia126_utt10.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3328 + }, + { + "path": "dia135_utt4.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3329 + }, + { + "path": "dia163_utt3.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3330 + }, + { + "path": "dia241_utt4.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3331 + }, + { + "path": "dia216_utt0.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3332 + }, + { + "path": "dia254_utt0.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3333 + }, + { + "path": "dia49_utt8.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3334 + }, + { + "path": "dia123_utt7.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3335 + }, + { + "path": "dia165_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3336 + }, + { + "path": "dia243_utt5.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3337 + }, + { + "path": "dia255_utt8.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3338 + }, + { + "path": "dia71_utt10.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3339 + }, + { + "path": "dia178_utt4.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3340 + }, + { + "path": "dia2_utt8.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3341 + }, + { + "path": "dia231_utt3.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3342 + }, + { + "path": "dia38_utt3.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3343 + }, + { + "path": "dia111_utt11.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3344 + }, + { + "path": "dia214_utt3.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3345 + }, + { + "path": "dia16_utt9.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3346 + }, + { + "path": "dia137_utt11.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3347 + }, + { + "path": "dia102_utt8.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3348 + }, + { + "path": "dia182_utt1.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3349 + }, + { + "path": "dia125_utt20.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3350 + }, + { + "path": "dia58_utt8.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3351 + }, + { + "path": "dia259_utt2.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3352 + }, + { + "path": "dia192_utt8.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3353 + }, + { + "path": "dia133_utt10.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3354 + }, + { + "path": "dia213_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3355 + }, + { + "path": "dia41_utt3.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3356 + }, + { + "path": "dia183_utt8.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3357 + }, + { + "path": "dia29_utt9.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3358 + }, + { + "path": "dia56_utt10.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3359 + }, + { + "path": "dia72_utt13.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3360 + }, + { + "path": "dia266_utt2.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3361 + }, + { + "path": "dia223_utt3.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3362 + }, + { + "path": "dia130_utt15.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3363 + }, + { + "path": "dia63_utt0.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3364 + }, + { + "path": "dia233_utt4.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3365 + }, + { + "path": "dia37_utt1.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3366 + }, + { + "path": "dia128_utt3.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3367 + }, + { + "path": "dia125_utt2.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3368 + }, + { + "path": "dia204_utt3.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3369 + }, + { + "path": "dia95_utt12.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3370 + }, + { + "path": "dia54_utt3.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3371 + }, + { + "path": "dia173_utt18.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3372 + }, + { + "path": "dia25_utt0.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3373 + }, + { + "path": "dia12_utt14.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3374 + }, + { + "path": "dia99_utt4.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3375 + }, + { + "path": "dia116_utt6.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3376 + }, + { + "path": "dia95_utt7.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3377 + }, + { + "path": "dia223_utt4.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3378 + }, + { + "path": "dia223_utt6.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3379 + }, + { + "path": "dia65_utt9.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3380 + }, + { + "path": "dia169_utt2.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3381 + }, + { + "path": "dia35_utt3.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3382 + }, + { + "path": "dia123_utt8.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3383 + }, + { + "path": "dia1_utt2.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3384 + }, + { + "path": "dia225_utt2.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3385 + }, + { + "path": "dia222_utt0.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3386 + }, + { + "path": "dia244_utt17.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3387 + }, + { + "path": "dia210_utt7.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3388 + }, + { + "path": "dia213_utt6.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3389 + }, + { + "path": "dia90_utt5.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3390 + }, + { + "path": "dia233_utt1.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3391 + }, + { + "path": "dia218_utt1.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3392 + }, + { + "path": "dia34_utt12.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3393 + }, + { + "path": "dia166_utt8.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3394 + }, + { + "path": "dia25_utt7.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3395 + }, + { + "path": "dia111_utt13.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3396 + }, + { + "path": "dia155_utt5.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3397 + }, + { + "path": "dia56_utt9.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3398 + }, + { + "path": "dia128_utt5.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3399 + }, + { + "path": "dia6_utt0.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3400 + }, + { + "path": "dia188_utt5.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3401 + }, + { + "path": "dia100_utt9.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3402 + }, + { + "path": "dia20_utt4.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3403 + }, + { + "path": "dia175_utt2.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3404 + }, + { + "path": "dia278_utt13.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3405 + }, + { + "path": "dia41_utt2.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3406 + }, + { + "path": "dia70_utt4.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3407 + }, + { + "path": "dia125_utt22.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3408 + }, + { + "path": "dia244_utt2.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3409 + }, + { + "path": "dia121_utt0.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3410 + }, + { + "path": "dia147_utt7.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3411 + }, + { + "path": "dia78_utt1.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3412 + }, + { + "path": "dia233_utt10.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3413 + }, + { + "path": "dia100_utt7.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3414 + }, + { + "path": "dia259_utt9.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3415 + }, + { + "path": "dia131_utt8.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3416 + }, + { + "path": "dia222_utt13.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3417 + }, + { + "path": "dia12_utt18.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3418 + }, + { + "path": "dia132_utt4.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3419 + }, + { + "path": "dia68_utt11.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3420 + }, + { + "path": "dia110_utt4.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3421 + }, + { + "path": "dia190_utt2.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3422 + }, + { + "path": "dia122_utt0.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3423 + }, + { + "path": "dia37_utt3.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3424 + }, + { + "path": "dia58_utt5.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3425 + }, + { + "path": "dia130_utt5.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3426 + }, + { + "path": "dia173_utt13.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3427 + }, + { + "path": "dia97_utt9.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3428 + }, + { + "path": "dia218_utt0.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3429 + }, + { + "path": "dia112_utt4.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3430 + }, + { + "path": "dia47_utt12.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3431 + }, + { + "path": "dia126_utt4.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3432 + }, + { + "path": "dia217_utt4.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3433 + }, + { + "path": "dia70_utt12.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3434 + }, + { + "path": "dia0_utt2.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3435 + }, + { + "path": "dia259_utt7.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3436 + }, + { + "path": "dia116_utt8.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3437 + }, + { + "path": "dia49_utt1.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3438 + }, + { + "path": "dia17_utt13.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3439 + }, + { + "path": "dia71_utt2.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3440 + }, + { + "path": "dia178_utt7.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3441 + }, + { + "path": "dia72_utt10.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3442 + }, + { + "path": "dia18_utt2.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3443 + }, + { + "path": "dia108_utt0.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3444 + }, + { + "path": "dia12_utt10.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3445 + }, + { + "path": "dia73_utt8.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3446 + }, + { + "path": "dia188_utt7.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3447 + }, + { + "path": "dia242_utt4.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3448 + }, + { + "path": "dia152_utt0.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3449 + }, + { + "path": "dia198_utt5.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3450 + }, + { + "path": "dia254_utt3.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3451 + }, + { + "path": "dia49_utt10.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3452 + }, + { + "path": "dia178_utt5.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3453 + }, + { + "path": "dia34_utt2.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3454 + }, + { + "path": "dia97_utt1.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3455 + }, + { + "path": "dia144_utt6.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3456 + }, + { + "path": "dia72_utt6.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3457 + }, + { + "path": "dia158_utt1.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3458 + }, + { + "path": "dia147_utt3.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3459 + }, + { + "path": "dia34_utt11.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3460 + }, + { + "path": "dia64_utt4.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3461 + }, + { + "path": "dia88_utt8.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3462 + }, + { + "path": "dia117_utt1.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3463 + }, + { + "path": "dia169_utt6.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3464 + }, + { + "path": "dia132_utt13.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3465 + }, + { + "path": "dia261_utt8.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3466 + }, + { + "path": "dia107_utt5.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3467 + }, + { + "path": "dia68_utt7.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3468 + }, + { + "path": "dia25_utt13.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3469 + }, + { + "path": "dia75_utt4.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3470 + }, + { + "path": "dia91_utt0.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3471 + }, + { + "path": "dia72_utt15.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3472 + }, + { + "path": "dia102_utt1.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3473 + }, + { + "path": "dia41_utt1.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3474 + }, + { + "path": "dia140_utt6.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3475 + }, + { + "path": "dia140_utt10.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3476 + }, + { + "path": "dia124_utt14.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3477 + }, + { + "path": "dia109_utt3.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3478 + }, + { + "path": "dia244_utt14.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3479 + }, + { + "path": "dia68_utt4.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3480 + }, + { + "path": "dia9_utt12.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3481 + }, + { + "path": "dia194_utt3.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3482 + }, + { + "path": "dia9_utt8.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3483 + }, + { + "path": "dia7_utt5.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3484 + }, + { + "path": "dia269_utt1.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3485 + }, + { + "path": "dia17_utt8.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3486 + }, + { + "path": "dia250_utt1.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3487 + }, + { + "path": "dia58_utt1.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3488 + }, + { + "path": "dia140_utt16.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3489 + }, + { + "path": "dia238_utt4.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3490 + }, + { + "path": "dia22_utt10.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3491 + }, + { + "path": "dia47_utt13.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3492 + }, + { + "path": "dia237_utt9.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3493 + }, + { + "path": "dia178_utt6.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3494 + }, + { + "path": "dia88_utt11.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3495 + }, + { + "path": "dia155_utt7.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3496 + }, + { + "path": "dia53_utt12.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3497 + }, + { + "path": "dia42_utt4.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3498 + }, + { + "path": "dia241_utt9.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3499 + }, + { + "path": "dia237_utt11.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3500 + }, + { + "path": "dia230_utt1.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3501 + }, + { + "path": "dia95_utt4.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3502 + }, + { + "path": "dia44_utt2.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3503 + }, + { + "path": "dia279_utt11.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3504 + }, + { + "path": "dia78_utt2.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3505 + }, + { + "path": "dia199_utt11.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3506 + }, + { + "path": "dia113_utt9.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3507 + }, + { + "path": "dia77_utt3.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3508 + }, + { + "path": "dia239_utt6.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3509 + }, + { + "path": "dia263_utt2.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3510 + }, + { + "path": "dia125_utt6.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3511 + }, + { + "path": "dia147_utt8.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3512 + }, + { + "path": "dia142_utt0.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3513 + }, + { + "path": "dia233_utt11.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3514 + }, + { + "path": "dia55_utt2.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3515 + }, + { + "path": "dia53_utt0.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3516 + }, + { + "path": "dia230_utt6.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3517 + }, + { + "path": "dia236_utt0.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3518 + }, + { + "path": "dia17_utt29.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3519 + }, + { + "path": "dia16_utt4.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3520 + }, + { + "path": "dia186_utt3.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3521 + }, + { + "path": "dia65_utt3.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3522 + }, + { + "path": "dia183_utt6.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3523 + }, + { + "path": "dia157_utt6.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3524 + }, + { + "path": "dia86_utt2.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3525 + }, + { + "path": "dia5_utt5.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3526 + }, + { + "path": "dia112_utt3.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3527 + }, + { + "path": "dia186_utt0.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3528 + }, + { + "path": "dia67_utt3.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3529 + }, + { + "path": "dia2_utt0.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3530 + }, + { + "path": "dia246_utt12.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3531 + }, + { + "path": "dia219_utt1.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3532 + }, + { + "path": "dia199_utt6.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3533 + }, + { + "path": "dia96_utt10.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3534 + }, + { + "path": "dia68_utt9.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3535 + }, + { + "path": "dia154_utt19.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3536 + }, + { + "path": "dia262_utt14.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3537 + }, + { + "path": "dia76_utt6.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3538 + }, + { + "path": "dia228_utt11.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3539 + }, + { + "path": "dia110_utt6.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3540 + }, + { + "path": "dia231_utt12.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3541 + }, + { + "path": "dia67_utt0.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3542 + }, + { + "path": "dia57_utt1.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3543 + }, + { + "path": "dia227_utt0.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3544 + }, + { + "path": "dia259_utt13.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3545 + }, + { + "path": "dia47_utt8.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3546 + }, + { + "path": "dia198_utt12.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3547 + }, + { + "path": "dia210_utt1.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3548 + }, + { + "path": "dia92_utt3.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3549 + }, + { + "path": "dia122_utt2.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3550 + }, + { + "path": "dia133_utt5.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3551 + }, + { + "path": "dia184_utt2.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3552 + }, + { + "path": "dia191_utt9.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3553 + }, + { + "path": "dia12_utt11.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3554 + }, + { + "path": "dia81_utt0.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3555 + }, + { + "path": "dia28_utt12.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3556 + }, + { + "path": "dia221_utt1.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3557 + }, + { + "path": "dia43_utt0.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3558 + }, + { + "path": "dia182_utt4.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3559 + }, + { + "path": "dia244_utt0.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3560 + }, + { + "path": "dia71_utt0.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3561 + }, + { + "path": "dia73_utt9.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3562 + }, + { + "path": "dia73_utt0.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3563 + }, + { + "path": "dia219_utt0.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3564 + }, + { + "path": "dia237_utt2.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3565 + }, + { + "path": "dia12_utt2.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3566 + }, + { + "path": "dia86_utt0.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3567 + }, + { + "path": "dia42_utt0.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3568 + }, + { + "path": "dia96_utt19.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3569 + }, + { + "path": "dia258_utt12.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3570 + }, + { + "path": "dia64_utt14.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3571 + }, + { + "path": "dia234_utt10.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3572 + }, + { + "path": "dia49_utt0.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3573 + }, + { + "path": "dia169_utt21.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3574 + }, + { + "path": "dia189_utt6.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3575 + }, + { + "path": "dia232_utt1.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3576 + }, + { + "path": "dia65_utt4.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3577 + }, + { + "path": "dia195_utt11.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3578 + }, + { + "path": "dia39_utt2.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3579 + }, + { + "path": "dia100_utt11.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3580 + }, + { + "path": "dia226_utt2.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3581 + }, + { + "path": "dia94_utt1.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3582 + }, + { + "path": "dia166_utt13.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3583 + }, + { + "path": "dia153_utt3.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3584 + }, + { + "path": "dia70_utt21.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3585 + }, + { + "path": "dia259_utt10.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3586 + }, + { + "path": "dia162_utt0.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3587 + }, + { + "path": "dia24_utt3.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3588 + }, + { + "path": "dia124_utt3.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3589 + }, + { + "path": "dia168_utt2.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3590 + }, + { + "path": "dia16_utt2.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3591 + }, + { + "path": "dia1_utt1.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3592 + }, + { + "path": "dia266_utt9.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3593 + }, + { + "path": "dia119_utt0.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3594 + }, + { + "path": "dia195_utt9.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3595 + }, + { + "path": "dia164_utt7.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3596 + }, + { + "path": "dia137_utt0.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3597 + }, + { + "path": "dia15_utt4.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3598 + }, + { + "path": "dia66_utt4.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3599 + }, + { + "path": "dia183_utt3.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3600 + }, + { + "path": "dia154_utt4.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3601 + }, + { + "path": "dia188_utt6.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3602 + }, + { + "path": "dia131_utt4.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3603 + }, + { + "path": "dia279_utt5.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3604 + }, + { + "path": "dia120_utt1.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3605 + }, + { + "path": "dia73_utt7.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3606 + }, + { + "path": "dia260_utt2.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3607 + }, + { + "path": "dia21_utt4.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3608 + }, + { + "path": "dia149_utt0.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3609 + }, + { + "path": "dia120_utt10.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3610 + }, + { + "path": "dia134_utt1.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3611 + }, + { + "path": "dia263_utt3.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3612 + }, + { + "path": "dia230_utt7.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3613 + }, + { + "path": "dia12_utt5.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3614 + }, + { + "path": "dia223_utt2.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3615 + }, + { + "path": "dia263_utt8.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3616 + }, + { + "path": "dia76_utt1.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3617 + }, + { + "path": "dia173_utt11.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3618 + }, + { + "path": "dia258_utt3.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3619 + }, + { + "path": "dia144_utt9.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3620 + }, + { + "path": "dia120_utt15.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3621 + }, + { + "path": "dia70_utt14.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3622 + }, + { + "path": "dia259_utt8.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3623 + }, + { + "path": "dia143_utt0.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3624 + }, + { + "path": "dia96_utt11.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3625 + }, + { + "path": "dia160_utt1.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3626 + }, + { + "path": "dia167_utt9.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3627 + }, + { + "path": "dia255_utt3.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3628 + }, + { + "path": "dia139_utt8.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3629 + }, + { + "path": "dia64_utt12.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3630 + }, + { + "path": "dia135_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3631 + }, + { + "path": "dia76_utt8.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3632 + }, + { + "path": "dia117_utt3.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3633 + }, + { + "path": "dia41_utt7.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3634 + }, + { + "path": "dia258_utt7.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3635 + }, + { + "path": "dia145_utt6.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3636 + }, + { + "path": "dia201_utt11.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3637 + }, + { + "path": "dia13_utt3.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3638 + }, + { + "path": "dia120_utt7.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3639 + }, + { + "path": "dia173_utt2.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3640 + }, + { + "path": "dia153_utt6.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3641 + }, + { + "path": "dia242_utt0.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3642 + }, + { + "path": "dia258_utt0.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3643 + }, + { + "path": "dia123_utt14.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3644 + }, + { + "path": "dia163_utt1.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3645 + }, + { + "path": "dia88_utt6.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3646 + }, + { + "path": "dia251_utt1.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3647 + }, + { + "path": "dia196_utt10.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3648 + }, + { + "path": "dia60_utt3.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3649 + }, + { + "path": "dia131_utt3.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3650 + }, + { + "path": "dia212_utt5.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3651 + }, + { + "path": "dia226_utt6.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3652 + }, + { + "path": "dia49_utt7.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3653 + }, + { + "path": "dia177_utt0.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3654 + }, + { + "path": "dia279_utt1.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3655 + }, + { + "path": "dia96_utt20.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3656 + }, + { + "path": "dia254_utt12.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3657 + }, + { + "path": "dia34_utt13.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3658 + }, + { + "path": "dia274_utt3.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3659 + }, + { + "path": "dia170_utt0.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3660 + }, + { + "path": "dia255_utt1.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3661 + }, + { + "path": "dia188_utt1.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3662 + }, + { + "path": "dia69_utt5.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3663 + }, + { + "path": "dia113_utt6.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3664 + }, + { + "path": "dia151_utt2.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3665 + }, + { + "path": "dia17_utt32.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3666 + }, + { + "path": "dia96_utt4.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3667 + }, + { + "path": "dia54_utt6.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3668 + }, + { + "path": "dia55_utt0.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3669 + }, + { + "path": "dia254_utt14.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3670 + }, + { + "path": "dia208_utt8.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3671 + }, + { + "path": "dia252_utt18.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3672 + }, + { + "path": "dia190_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3673 + }, + { + "path": "dia137_utt5.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3674 + }, + { + "path": "dia93_utt10.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3675 + }, + { + "path": "dia113_utt4.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3676 + }, + { + "path": "dia198_utt4.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3677 + }, + { + "path": "dia119_utt5.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3678 + }, + { + "path": "dia270_utt12.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3679 + }, + { + "path": "dia173_utt14.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3680 + }, + { + "path": "dia39_utt4.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3681 + }, + { + "path": "dia158_utt5.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3682 + }, + { + "path": "dia65_utt7.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3683 + }, + { + "path": "dia9_utt4.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3684 + }, + { + "path": "dia2_utt4.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3685 + }, + { + "path": "dia96_utt13.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3686 + }, + { + "path": "dia174_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3687 + }, + { + "path": "dia1_utt3.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3688 + }, + { + "path": "dia96_utt9.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3689 + }, + { + "path": "dia254_utt4.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3690 + }, + { + "path": "dia12_utt6.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3691 + }, + { + "path": "dia274_utt2.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3692 + }, + { + "path": "dia115_utt4.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3693 + }, + { + "path": "dia185_utt4.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3694 + }, + { + "path": "dia90_utt0.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3695 + }, + { + "path": "dia115_utt3.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3696 + }, + { + "path": "dia69_utt1.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3697 + }, + { + "path": "dia183_utt5.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3698 + }, + { + "path": "dia139_utt3.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3699 + }, + { + "path": "dia169_utt9.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3700 + }, + { + "path": "dia122_utt3.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3701 + }, + { + "path": "dia102_utt4.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3702 + }, + { + "path": "dia78_utt5.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3703 + }, + { + "path": "dia228_utt6.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3704 + }, + { + "path": "dia259_utt0.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3705 + }, + { + "path": "dia10_utt2.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3706 + }, + { + "path": "dia222_utt7.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3707 + }, + { + "path": "dia68_utt2.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3708 + }, + { + "path": "dia192_utt9.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3709 + }, + { + "path": "dia132_utt11.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3710 + }, + { + "path": "dia17_utt23.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3711 + }, + { + "path": "dia25_utt3.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3712 + }, + { + "path": "dia73_utt1.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3713 + }, + { + "path": "dia250_utt6.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3714 + }, + { + "path": "dia68_utt14.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3715 + }, + { + "path": "dia54_utt1.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3716 + }, + { + "path": "dia168_utt18.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3717 + }, + { + "path": "dia153_utt11.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3718 + }, + { + "path": "dia147_utt10.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3719 + }, + { + "path": "dia175_utt10.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3720 + }, + { + "path": "dia260_utt7.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3721 + }, + { + "path": "dia223_utt0.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3722 + }, + { + "path": "dia196_utt16.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3723 + }, + { + "path": "dia34_utt4.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3724 + }, + { + "path": "dia178_utt9.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3725 + }, + { + "path": "dia208_utt1.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3726 + }, + { + "path": "dia58_utt10.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3727 + }, + { + "path": "dia256_utt0.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3728 + }, + { + "path": "dia6_utt2.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3729 + }, + { + "path": "dia278_utt7.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3730 + }, + { + "path": "dia267_utt8.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3731 + }, + { + "path": "dia80_utt6.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3732 + }, + { + "path": "dia95_utt6.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3733 + }, + { + "path": "dia180_utt9.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3734 + }, + { + "path": "dia48_utt5.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3735 + }, + { + "path": "dia74_utt1.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3736 + }, + { + "path": "dia262_utt1.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3737 + }, + { + "path": "dia143_utt5.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3738 + }, + { + "path": "dia231_utt9.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3739 + }, + { + "path": "dia260_utt8.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3740 + }, + { + "path": "dia222_utt11.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3741 + }, + { + "path": "dia103_utt2.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3742 + }, + { + "path": "dia230_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3743 + }, + { + "path": "dia100_utt16.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3744 + }, + { + "path": "dia157_utt7.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3745 + }, + { + "path": "dia96_utt2.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3746 + }, + { + "path": "dia222_utt3.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3747 + }, + { + "path": "dia244_utt16.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3748 + }, + { + "path": "dia53_utt11.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3749 + }, + { + "path": "dia262_utt16.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3750 + }, + { + "path": "dia100_utt4.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3751 + }, + { + "path": "dia22_utt11.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3752 + }, + { + "path": "dia100_utt17.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3753 + }, + { + "path": "dia160_utt3.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3754 + }, + { + "path": "dia103_utt5.wav", + "question": "discern the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3755 + }, + { + "path": "dia25_utt15.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3756 + }, + { + "path": "dia118_utt9.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3757 + }, + { + "path": "dia187_utt8.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3758 + }, + { + "path": "dia274_utt4.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3759 + }, + { + "path": "dia161_utt1.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3760 + }, + { + "path": "dia49_utt3.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3761 + }, + { + "path": "dia17_utt9.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3762 + }, + { + "path": "dia173_utt3.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3763 + }, + { + "path": "dia185_utt7.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3764 + }, + { + "path": "dia167_utt1.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3765 + }, + { + "path": "dia29_utt2.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3766 + }, + { + "path": "dia83_utt9.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3767 + }, + { + "path": "dia198_utt0.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3768 + }, + { + "path": "dia205_utt7.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3769 + }, + { + "path": "dia58_utt13.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3770 + }, + { + "path": "dia17_utt3.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3771 + }, + { + "path": "dia134_utt3.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3772 + }, + { + "path": "dia32_utt4.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3773 + }, + { + "path": "dia258_utt5.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3774 + }, + { + "path": "dia93_utt9.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3775 + }, + { + "path": "dia170_utt4.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3776 + }, + { + "path": "dia182_utt6.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3777 + }, + { + "path": "dia112_utt13.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3778 + }, + { + "path": "dia278_utt11.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3779 + }, + { + "path": "dia54_utt7.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3780 + }, + { + "path": "dia130_utt1.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3781 + }, + { + "path": "dia258_utt11.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3782 + }, + { + "path": "dia120_utt16.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3783 + }, + { + "path": "dia68_utt8.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3784 + }, + { + "path": "dia188_utt2.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3785 + }, + { + "path": "dia220_utt0.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3786 + }, + { + "path": "dia145_utt11.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3787 + }, + { + "path": "dia4_utt1.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3788 + }, + { + "path": "dia7_utt0.wav", + "question": "identify the gender of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3789 + }, + { + "path": "dia263_utt6.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3790 + }, + { + "path": "dia15_utt2.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3791 + }, + { + "path": "dia70_utt0.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3792 + }, + { + "path": "dia78_utt6.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3793 + }, + { + "path": "dia42_utt2.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3794 + }, + { + "path": "dia208_utt3.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3795 + }, + { + "path": "dia195_utt1.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3796 + }, + { + "path": "dia73_utt10.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3797 + }, + { + "path": "dia153_utt2.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3798 + }, + { + "path": "dia117_utt4.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3799 + }, + { + "path": "dia54_utt11.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3800 + }, + { + "path": "dia140_utt8.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3801 + }, + { + "path": "dia83_utt6.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3802 + }, + { + "path": "dia262_utt7.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3803 + }, + { + "path": "dia72_utt7.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3804 + }, + { + "path": "dia49_utt9.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3805 + }, + { + "path": "dia115_utt1.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3806 + }, + { + "path": "dia120_utt2.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3807 + }, + { + "path": "dia270_utt10.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3808 + }, + { + "path": "dia80_utt3.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3809 + }, + { + "path": "dia85_utt0.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3810 + }, + { + "path": "dia166_utt5.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3811 + }, + { + "path": "dia246_utt0.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3812 + }, + { + "path": "dia25_utt1.wav", + "question": "What gender would you attribute to the speaker in this sound clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3813 + }, + { + "path": "dia252_utt11.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3814 + }, + { + "path": "dia7_utt3.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3815 + }, + { + "path": "dia249_utt1.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3816 + }, + { + "path": "dia79_utt0.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3817 + }, + { + "path": "dia208_utt4.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3818 + }, + { + "path": "dia64_utt16.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3819 + }, + { + "path": "dia41_utt9.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3820 + }, + { + "path": "dia146_utt2.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3821 + }, + { + "path": "dia168_utt0.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3822 + }, + { + "path": "dia221_utt4.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3823 + }, + { + "path": "dia189_utt7.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3824 + }, + { + "path": "dia106_utt0.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3825 + }, + { + "path": "dia229_utt7.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3826 + }, + { + "path": "dia4_utt5.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3827 + }, + { + "path": "dia121_utt2.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3828 + }, + { + "path": "dia114_utt0.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3829 + }, + { + "path": "dia10_utt0.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3830 + }, + { + "path": "dia34_utt7.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3831 + }, + { + "path": "dia102_utt2.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3832 + }, + { + "path": "dia222_utt10.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3833 + }, + { + "path": "dia189_utt9.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3834 + }, + { + "path": "dia116_utt5.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3835 + }, + { + "path": "dia111_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3836 + }, + { + "path": "dia60_utt2.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3837 + }, + { + "path": "dia107_utt3.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3838 + }, + { + "path": "dia189_utt2.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3839 + }, + { + "path": "dia269_utt2.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3840 + }, + { + "path": "dia1_utt4.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3841 + }, + { + "path": "dia244_utt21.wav", + "question": "make a determination about the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3842 + }, + { + "path": "dia217_utt6.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3843 + }, + { + "path": "dia98_utt2.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3844 + }, + { + "path": "dia100_utt6.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3845 + }, + { + "path": "dia233_utt6.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3846 + }, + { + "path": "dia4_utt2.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3847 + }, + { + "path": "dia259_utt1.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3848 + }, + { + "path": "dia16_utt7.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3849 + }, + { + "path": "dia208_utt9.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3850 + }, + { + "path": "dia34_utt10.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3851 + }, + { + "path": "dia258_utt4.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3852 + }, + { + "path": "dia250_utt3.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3853 + }, + { + "path": "dia45_utt7.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3854 + }, + { + "path": "dia196_utt2.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3855 + }, + { + "path": "dia202_utt5.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3856 + }, + { + "path": "dia76_utt4.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3857 + }, + { + "path": "dia110_utt5.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3858 + }, + { + "path": "dia28_utt16.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3859 + }, + { + "path": "dia87_utt1.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3860 + }, + { + "path": "dia196_utt9.wav", + "question": "detect the gender of the speaker in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3861 + }, + { + "path": "dia90_utt4.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3862 + }, + { + "path": "dia167_utt3.wav", + "question": "tell the gender of the speaker from this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3863 + }, + { + "path": "dia80_utt2.wav", + "question": "tell the gender of the speaker in this sound file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3864 + }, + { + "path": "dia184_utt1.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3865 + }, + { + "path": "dia69_utt0.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3866 + }, + { + "path": "dia25_utt9.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3867 + }, + { + "path": "dia199_utt5.wav", + "question": "identify the gender of the speaker in this sound recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3868 + }, + { + "path": "dia161_utt7.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3869 + }, + { + "path": "dia28_utt6.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3870 + }, + { + "path": "dia152_utt4.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3871 + }, + { + "path": "dia278_utt5.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3872 + }, + { + "path": "dia233_utt9.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3873 + }, + { + "path": "dia90_utt3.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3874 + }, + { + "path": "dia17_utt22.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3875 + }, + { + "path": "dia22_utt15.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3876 + }, + { + "path": "dia174_utt5.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3877 + }, + { + "path": "dia131_utt6.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3878 + }, + { + "path": "dia78_utt8.wav", + "question": "establish the gender of the person whose voice is recorded here.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3879 + }, + { + "path": "dia22_utt0.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3880 + }, + { + "path": "dia120_utt8.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3881 + }, + { + "path": "dia153_utt9.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3882 + }, + { + "path": "dia100_utt3.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3883 + }, + { + "path": "dia89_utt2.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3884 + }, + { + "path": "dia170_utt1.wav", + "question": "What is the sex of the individual speaking in this audio?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3885 + }, + { + "path": "dia45_utt4.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3886 + }, + { + "path": "dia138_utt3.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3887 + }, + { + "path": "dia165_utt3.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3888 + }, + { + "path": "dia160_utt11.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3889 + }, + { + "path": "dia273_utt1.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3890 + }, + { + "path": "dia71_utt12.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3891 + }, + { + "path": "dia147_utt0.wav", + "question": "determine the sex of the speaker in this sound file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3892 + }, + { + "path": "dia113_utt12.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3893 + }, + { + "path": "dia146_utt3.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3894 + }, + { + "path": "dia9_utt10.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3895 + }, + { + "path": "dia213_utt2.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3896 + }, + { + "path": "dia154_utt3.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3897 + }, + { + "path": "dia156_utt1.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3898 + }, + { + "path": "dia126_utt8.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3899 + }, + { + "path": "dia137_utt1.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3900 + }, + { + "path": "dia100_utt5.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3901 + }, + { + "path": "dia113_utt3.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3902 + }, + { + "path": "dia122_utt1.wav", + "question": "deduce the gender of the person speaking in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3903 + }, + { + "path": "dia204_utt6.wav", + "question": "infer the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3904 + }, + { + "path": "dia241_utt2.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3905 + }, + { + "path": "dia76_utt16.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3906 + }, + { + "path": "dia155_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3907 + }, + { + "path": "dia68_utt13.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3908 + }, + { + "path": "dia124_utt9.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3909 + }, + { + "path": "dia175_utt6.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3910 + }, + { + "path": "dia22_utt6.wav", + "question": "discern the gender of the individual in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3911 + }, + { + "path": "dia251_utt3.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3912 + }, + { + "path": "dia21_utt6.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3913 + }, + { + "path": "dia17_utt15.wav", + "question": "What is the gender identity of the speaker in this audio clip?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3914 + }, + { + "path": "dia125_utt9.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3915 + }, + { + "path": "dia213_utt5.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3916 + }, + { + "path": "dia153_utt10.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3917 + }, + { + "path": "dia98_utt9.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3918 + }, + { + "path": "dia182_utt5.wav", + "question": "figure out the gender of the voice heard in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3919 + }, + { + "path": "dia144_utt10.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3920 + }, + { + "path": "dia270_utt2.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3921 + }, + { + "path": "dia244_utt3.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3922 + }, + { + "path": "dia68_utt1.wav", + "question": "By listening to this audio, guess the gender of the speaker.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3923 + }, + { + "path": "dia229_utt0.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3924 + }, + { + "path": "dia196_utt5.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3925 + }, + { + "path": "dia61_utt1.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3926 + }, + { + "path": "dia17_utt12.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3927 + }, + { + "path": "dia14_utt3.wav", + "question": "tell me the gender of the individual speaking in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3928 + }, + { + "path": "dia108_utt7.wav", + "question": "ascertain the sex of the voice in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3929 + }, + { + "path": "dia226_utt8.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3930 + }, + { + "path": "dia111_utt12.wav", + "question": "identify the sex of the individual speaking in this audio clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3931 + }, + { + "path": "dia257_utt3.wav", + "question": "please identify the gender of the speaker in this audio recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3932 + }, + { + "path": "dia230_utt8.wav", + "question": "tell me the gender of the person we hear in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3933 + }, + { + "path": "dia70_utt16.wav", + "question": "detect the gender of the voice present in this audio clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3934 + }, + { + "path": "dia144_utt0.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3935 + }, + { + "path": "dia195_utt7.wav", + "question": "recognize the gender of the person in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3936 + }, + { + "path": "dia99_utt9.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3937 + }, + { + "path": "dia243_utt3.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3938 + }, + { + "path": "dia196_utt4.wav", + "question": "detect the gender of the person whose voice is on this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3939 + }, + { + "path": "dia228_utt12.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3940 + }, + { + "path": "dia33_utt8.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3941 + }, + { + "path": "dia161_utt11.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3942 + }, + { + "path": "dia56_utt5.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3943 + }, + { + "path": "dia33_utt1.wav", + "question": "determine the gender of the voice in this audio file.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3944 + }, + { + "path": "dia34_utt5.wav", + "question": "identify the speaker's gender from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3945 + }, + { + "path": "dia95_utt19.wav", + "question": "determining the sex of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3946 + }, + { + "path": "dia123_utt2.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3947 + }, + { + "path": "dia240_utt4.wav", + "question": "pinpoint the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3948 + }, + { + "path": "dia263_utt5.wav", + "question": "determine the speaker's gender in this sound recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3949 + }, + { + "path": "dia53_utt1.wav", + "question": "identify the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3950 + }, + { + "path": "dia254_utt2.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3951 + }, + { + "path": "dia95_utt2.wav", + "question": "What gender does the speaker in the audio identify as?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3952 + }, + { + "path": "dia174_utt17.wav", + "question": "make out the gender of the individual speaking in this audio track.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3953 + }, + { + "path": "dia48_utt1.wav", + "question": "What is your perception of the speaker's gender in this audio piece?", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3954 + }, + { + "path": "dia193_utt3.wav", + "question": "determine the speaker's gender in this audio recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3955 + }, + { + "path": "dia12_utt4.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3956 + }, + { + "path": "dia99_utt7.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3957 + }, + { + "path": "dia239_utt2.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3958 + }, + { + "path": "dia271_utt0.wav", + "question": "what gender the speaker is in this audio recording?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3959 + }, + { + "path": "dia22_utt1.wav", + "question": "figure out the speaker's gender in this sound clip.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3960 + }, + { + "path": "dia96_utt18.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3961 + }, + { + "path": "dia228_utt5.wav", + "question": "figure out the gender of the speaker from this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3962 + }, + { + "path": "dia223_utt1.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3963 + }, + { + "path": "dia34_utt17.wav", + "question": "tell the gender of the person from their voice in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Woman", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3964 + }, + { + "path": "dia12_utt7.wav", + "question": "classify the speaker's gender in this audio file.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3965 + }, + { + "path": "dia230_utt4.wav", + "question": "identify the gender of the person speaking in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3966 + }, + { + "path": "dia65_utt1.wav", + "question": "ascertain the gender of the voice in this recording.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Female", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3967 + }, + { + "path": "dia275_utt1.wav", + "question": "tell the gender of the speaker in this audio.", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3968 + }, + { + "path": "dia83_utt5.wav", + "question": "perceive the gender of the individual whose voice is in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3969 + }, + { + "path": "dia118_utt2.wav", + "question": "Regarding the audio's speaker, what gender do they represent?", + "choice_a": "Male", + "choice_b": "Female", + "answer_gt": "Male", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3970 + }, + { + "path": "dia1_utt7.wav", + "question": "make out the gender of the speaker in this recording.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3971 + }, + { + "path": "dia29_utt8.wav", + "question": "specify the speaker's gender based on their voice in this audio.", + "choice_a": "Man", + "choice_b": "Woman", + "answer_gt": "Man", + "task_name": "Speaker_Gender_Recognition", + "dataset_name": "meld", + "uniq_id": 3972 + }, + { + "path": "Ses01M_impro05_F025.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3973 + }, + { + "path": "Ses01F_impro04_F033.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3974 + }, + { + "path": "Ses05F_impro05_F034.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3975 + }, + { + "path": "Ses05M_impro03_M010.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3976 + }, + { + "path": "Ses05M_script02_1_M031.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3977 + }, + { + "path": "Ses05F_script02_1_M018.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3978 + }, + { + "path": "Ses02M_impro05_M001.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3979 + }, + { + "path": "Ses03M_impro05b_F001.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3980 + }, + { + "path": "Ses03M_impro05a_M000.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3981 + }, + { + "path": "Ses05F_script03_1_M025.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3982 + }, + { + "path": "Ses01F_impro05_M026.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3983 + }, + { + "path": "Ses02M_script02_1_M004.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3984 + }, + { + "path": "Ses01F_impro01_F005.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3985 + }, + { + "path": "Ses01F_script02_1_M031.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3986 + }, + { + "path": "Ses02M_impro02_F012.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3987 + }, + { + "path": "Ses01M_script01_3_M040.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3988 + }, + { + "path": "Ses04M_script03_1_F014.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3989 + }, + { + "path": "Ses01F_impro07_M005.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3990 + }, + { + "path": "Ses03M_script01_3_M037.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3991 + }, + { + "path": "Ses03F_script02_2_F020.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3992 + }, + { + "path": "Ses03M_script02_1_M027.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3993 + }, + { + "path": "Ses02M_impro07_F018.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3994 + }, + { + "path": "Ses05M_script03_2_F038.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3995 + }, + { + "path": "Ses01F_script03_2_M004.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3996 + }, + { + "path": "Ses02F_script03_2_M041.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3997 + }, + { + "path": "Ses05M_impro02_M011.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3998 + }, + { + "path": "Ses02F_impro03_F000.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 3999 + }, + { + "path": "Ses03M_impro08a_F021.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4000 + }, + { + "path": "Ses05F_impro06_F022.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4001 + }, + { + "path": "Ses05F_impro05_M017.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4002 + }, + { + "path": "Ses05M_script03_1_M022.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4003 + }, + { + "path": "Ses05M_script03_1_M025.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4004 + }, + { + "path": "Ses03F_impro03_F009.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4005 + }, + { + "path": "Ses03M_script01_1_M041.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4006 + }, + { + "path": "Ses02M_impro05_M026.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4007 + }, + { + "path": "Ses04F_script02_1_M034.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4008 + }, + { + "path": "Ses05F_script03_1_M016.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4009 + }, + { + "path": "Ses03F_impro02_F037.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4010 + }, + { + "path": "Ses05F_script01_1_F026.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4011 + }, + { + "path": "Ses01F_impro05_M030.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4012 + }, + { + "path": "Ses02M_script02_1_M029.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4013 + }, + { + "path": "Ses02M_impro08_F000.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4014 + }, + { + "path": "Ses04M_impro08_F028.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4015 + }, + { + "path": "Ses04M_impro05_M022.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4016 + }, + { + "path": "Ses03F_script01_2_F016.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4017 + }, + { + "path": "Ses04F_impro06_F002.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4018 + }, + { + "path": "Ses03F_script02_2_F012.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4019 + }, + { + "path": "Ses04F_script02_2_M021.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4020 + }, + { + "path": "Ses04F_impro02_M021.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4021 + }, + { + "path": "Ses04F_impro05_F012.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4022 + }, + { + "path": "Ses02F_impro08_M007.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4023 + }, + { + "path": "Ses04M_impro04_M021.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4024 + }, + { + "path": "Ses02F_impro06_M019.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4025 + }, + { + "path": "Ses05F_script01_1_F031.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4026 + }, + { + "path": "Ses03M_script02_2_M037.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4027 + }, + { + "path": "Ses02M_impro02_M008.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4028 + }, + { + "path": "Ses03F_impro08_F002.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4029 + }, + { + "path": "Ses01F_impro02_F010.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4030 + }, + { + "path": "Ses01M_script03_1_M039.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4031 + }, + { + "path": "Ses05M_script01_1_M004.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4032 + }, + { + "path": "Ses03F_impro06_F023.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4033 + }, + { + "path": "Ses01F_impro06_M001.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4034 + }, + { + "path": "Ses01M_script01_3_M034.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4035 + }, + { + "path": "Ses01M_impro07_M008.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4036 + }, + { + "path": "Ses05F_impro03_M039.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4037 + }, + { + "path": "Ses02F_script03_1_M029.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4038 + }, + { + "path": "Ses04F_script03_1_F017.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4039 + }, + { + "path": "Ses04M_impro03_M029.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4040 + }, + { + "path": "Ses04F_impro03_M033.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4041 + }, + { + "path": "Ses05F_script01_3_M037.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4042 + }, + { + "path": "Ses03F_script02_2_F035.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4043 + }, + { + "path": "Ses04M_script02_2_M031.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4044 + }, + { + "path": "Ses02M_impro03_M020.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4045 + }, + { + "path": "Ses03M_impro06_M018.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4046 + }, + { + "path": "Ses02M_impro04_F006.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4047 + }, + { + "path": "Ses03F_impro04_F022.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4048 + }, + { + "path": "Ses03F_script01_3_M041.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4049 + }, + { + "path": "Ses04F_script01_1_F032.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4050 + }, + { + "path": "Ses03M_script02_1_M023.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4051 + }, + { + "path": "Ses03M_impro08a_M018.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4052 + }, + { + "path": "Ses05M_impro08_M026.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4053 + }, + { + "path": "Ses02M_impro07_F028.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4054 + }, + { + "path": "Ses01M_script02_2_F035.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4055 + }, + { + "path": "Ses02F_impro07_M028.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4056 + }, + { + "path": "Ses02F_script02_2_F038.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4057 + }, + { + "path": "Ses03M_impro08b_M011.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4058 + }, + { + "path": "Ses02F_impro02_F008.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4059 + }, + { + "path": "Ses03F_script01_3_M040.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4060 + }, + { + "path": "Ses03M_script03_1_F005.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4061 + }, + { + "path": "Ses04F_impro05_F011.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4062 + }, + { + "path": "Ses02M_script01_3_M031.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4063 + }, + { + "path": "Ses01F_impro07_M012.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4064 + }, + { + "path": "Ses02M_script02_2_F031.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4065 + }, + { + "path": "Ses02M_script03_2_M024.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4066 + }, + { + "path": "Ses05F_script01_3_M034.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4067 + }, + { + "path": "Ses01M_impro03_F015.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4068 + }, + { + "path": "Ses01M_impro07_F016.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4069 + }, + { + "path": "Ses01M_script02_1_M045.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4070 + }, + { + "path": "Ses05M_script03_2_F037.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4071 + }, + { + "path": "Ses03F_script02_2_F043.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4072 + }, + { + "path": "Ses02M_script01_3_M030.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4073 + }, + { + "path": "Ses01M_script01_3_F003.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4074 + }, + { + "path": "Ses05M_script01_1_F014.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4075 + }, + { + "path": "Ses02F_script03_1_F017.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4076 + }, + { + "path": "Ses01M_script01_1_M035.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4077 + }, + { + "path": "Ses05F_script01_1_F038.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4078 + }, + { + "path": "Ses01F_script03_2_M002.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4079 + }, + { + "path": "Ses01F_script01_3_M032.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4080 + }, + { + "path": "Ses05F_impro05_M007.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4081 + }, + { + "path": "Ses05M_script02_2_F031.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4082 + }, + { + "path": "Ses02F_impro05_M016.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4083 + }, + { + "path": "Ses01M_script02_2_M043.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4084 + }, + { + "path": "Ses05M_script01_3_F019.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4085 + }, + { + "path": "Ses04M_impro07_M000.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4086 + }, + { + "path": "Ses05M_impro06_M002.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4087 + }, + { + "path": "Ses04F_script01_1_M028.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4088 + }, + { + "path": "Ses05F_impro05_F021.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4089 + }, + { + "path": "Ses05F_script01_2_F012.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4090 + }, + { + "path": "Ses05M_script01_1b_F028.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4091 + }, + { + "path": "Ses03M_impro03_M028.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4092 + }, + { + "path": "Ses01M_script01_3_M032.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4093 + }, + { + "path": "Ses01M_impro03_F005.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4094 + }, + { + "path": "Ses01F_script02_1_M026.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4095 + }, + { + "path": "Ses03M_impro05a_F012.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4096 + }, + { + "path": "Ses05F_impro07_M007.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4097 + }, + { + "path": "Ses03M_impro04_F023.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4098 + }, + { + "path": "Ses04F_impro06_F004.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4099 + }, + { + "path": "Ses02M_impro03_F013.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4100 + }, + { + "path": "Ses04F_script02_1_M032.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4101 + }, + { + "path": "Ses05M_impro06_M012.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4102 + }, + { + "path": "Ses03F_script01_1_M031.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4103 + }, + { + "path": "Ses02M_script01_3_F009.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4104 + }, + { + "path": "Ses02F_script02_1_M041.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4105 + }, + { + "path": "Ses01M_impro05_M023.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4106 + }, + { + "path": "Ses04F_impro03_M036.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4107 + }, + { + "path": "Ses01F_impro06_F028.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4108 + }, + { + "path": "Ses04M_impro07_M026.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4109 + }, + { + "path": "Ses05F_script03_2_F035.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4110 + }, + { + "path": "Ses03F_script01_1_M042.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4111 + }, + { + "path": "Ses05F_impro05_M010.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4112 + }, + { + "path": "Ses01F_impro06_M003.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4113 + }, + { + "path": "Ses04F_script01_3_F015.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4114 + }, + { + "path": "Ses03M_script01_3_F008.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4115 + }, + { + "path": "Ses04F_impro01_F014.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4116 + }, + { + "path": "Ses05F_script03_1_M033.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4117 + }, + { + "path": "Ses01F_impro05_F020.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4118 + }, + { + "path": "Ses05F_impro04_M008.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4119 + }, + { + "path": "Ses04F_impro03_F034.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4120 + }, + { + "path": "Ses03M_script02_2_F034.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4121 + }, + { + "path": "Ses02F_script02_2_M002.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4122 + }, + { + "path": "Ses01M_script03_2_M040.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4123 + }, + { + "path": "Ses02F_impro03_F011.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4124 + }, + { + "path": "Ses04M_impro03_M024.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4125 + }, + { + "path": "Ses03M_impro05b_F021.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4126 + }, + { + "path": "Ses05F_script03_2_F038.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4127 + }, + { + "path": "Ses01M_impro06_M028.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4128 + }, + { + "path": "Ses03F_impro03_F008.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4129 + }, + { + "path": "Ses01F_script03_2_F015.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4130 + }, + { + "path": "Ses01F_script03_2_M027.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4131 + }, + { + "path": "Ses05M_impro08_F027.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4132 + }, + { + "path": "Ses05F_impro05_M025.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4133 + }, + { + "path": "Ses03M_impro05b_M029.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4134 + }, + { + "path": "Ses01F_script01_1_M025.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4135 + }, + { + "path": "Ses03F_script02_2_F050.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4136 + }, + { + "path": "Ses01M_script03_2_F028.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4137 + }, + { + "path": "Ses01F_script03_1_M025.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4138 + }, + { + "path": "Ses02M_impro03_F023.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4139 + }, + { + "path": "Ses02F_impro04_M009.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4140 + }, + { + "path": "Ses05M_script01_3_M019.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4141 + }, + { + "path": "Ses05F_script03_1_M019.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4142 + }, + { + "path": "Ses03M_script01_3_F033.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4143 + }, + { + "path": "Ses04F_impro04_F018.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4144 + }, + { + "path": "Ses04F_script02_1_M009.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4145 + }, + { + "path": "Ses05F_script02_1_M011.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4146 + }, + { + "path": "Ses02M_impro05_M013.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4147 + }, + { + "path": "Ses04M_impro07_M010.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4148 + }, + { + "path": "Ses05F_impro07_M016.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4149 + }, + { + "path": "Ses05F_impro03_F057.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4150 + }, + { + "path": "Ses03M_impro03_M011.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4151 + }, + { + "path": "Ses03M_impro07_F006.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4152 + }, + { + "path": "Ses01F_impro04_M023.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4153 + }, + { + "path": "Ses02M_impro02_M009.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4154 + }, + { + "path": "Ses04F_impro03_M001.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4155 + }, + { + "path": "Ses01F_impro04_F001.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4156 + }, + { + "path": "Ses02F_script02_2_F043.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4157 + }, + { + "path": "Ses04M_script01_1_M022.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4158 + }, + { + "path": "Ses03M_impro06_F000.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4159 + }, + { + "path": "Ses05F_impro05_F010.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4160 + }, + { + "path": "Ses04F_impro07_F074.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4161 + }, + { + "path": "Ses05F_script01_3_M000.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4162 + }, + { + "path": "Ses02M_impro03_M027.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4163 + }, + { + "path": "Ses03F_script02_1_M042.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4164 + }, + { + "path": "Ses04M_script02_1_F018.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4165 + }, + { + "path": "Ses05M_impro02_M019.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4166 + }, + { + "path": "Ses02F_impro06_M011.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4167 + }, + { + "path": "Ses04F_impro07_M026.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4168 + }, + { + "path": "Ses02F_impro03_M015.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4169 + }, + { + "path": "Ses04M_impro06_M016.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4170 + }, + { + "path": "Ses04F_impro06_M007.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4171 + }, + { + "path": "Ses05F_impro06_F009.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4172 + }, + { + "path": "Ses04M_impro04_F015.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4173 + }, + { + "path": "Ses02M_script01_1_M003.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4174 + }, + { + "path": "Ses01F_script02_2_M016.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4175 + }, + { + "path": "Ses05M_impro03_M015.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4176 + }, + { + "path": "Ses01M_script01_3_M030.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4177 + }, + { + "path": "Ses03F_impro06_F005.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4178 + }, + { + "path": "Ses05M_impro02_F021.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4179 + }, + { + "path": "Ses05M_script01_1b_M002.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4180 + }, + { + "path": "Ses01M_impro03_M004.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4181 + }, + { + "path": "Ses02M_impro08_F007.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4182 + }, + { + "path": "Ses05F_script03_2_M013.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4183 + }, + { + "path": "Ses04M_script02_1_M018.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4184 + }, + { + "path": "Ses03M_impro05b_F016.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4185 + }, + { + "path": "Ses02M_script03_2_F003.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4186 + }, + { + "path": "Ses05M_script03_1_F024.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4187 + }, + { + "path": "Ses01F_script02_2_M006.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4188 + }, + { + "path": "Ses05F_impro05_M050.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4189 + }, + { + "path": "Ses01F_script01_2_F010.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4190 + }, + { + "path": "Ses05F_impro05_M006.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4191 + }, + { + "path": "Ses01F_impro02_M012.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4192 + }, + { + "path": "Ses03M_impro08a_M019.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4193 + }, + { + "path": "Ses05F_impro05_F020.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4194 + }, + { + "path": "Ses05F_impro06_F003.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4195 + }, + { + "path": "Ses01F_script01_3_M025.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4196 + }, + { + "path": "Ses05M_impro08_F011.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4197 + }, + { + "path": "Ses01F_script02_2_F009.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4198 + }, + { + "path": "Ses01F_script03_2_M031.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4199 + }, + { + "path": "Ses01M_script01_3_F005.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4200 + }, + { + "path": "Ses04F_script01_3_M026.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4201 + }, + { + "path": "Ses05F_impro02_F002.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4202 + }, + { + "path": "Ses04F_impro04_M017.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4203 + }, + { + "path": "Ses02M_script01_1_M028.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4204 + }, + { + "path": "Ses03M_impro07_M025.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4205 + }, + { + "path": "Ses01F_script01_3_M036.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4206 + }, + { + "path": "Ses02F_script03_2_F007.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4207 + }, + { + "path": "Ses01M_script03_2_F038.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4208 + }, + { + "path": "Ses03M_script01_2_F017.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4209 + }, + { + "path": "Ses02M_impro05_F015.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4210 + }, + { + "path": "Ses05F_script01_3_M009.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4211 + }, + { + "path": "Ses01F_impro01_M011.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4212 + }, + { + "path": "Ses03F_script01_1_M021.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4213 + }, + { + "path": "Ses03F_impro08_M011.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4214 + }, + { + "path": "Ses04M_script01_2_M019.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4215 + }, + { + "path": "Ses04M_impro05_F006.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4216 + }, + { + "path": "Ses02M_script03_2_F038.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4217 + }, + { + "path": "Ses05F_script02_2_F034.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4218 + }, + { + "path": "Ses05F_impro02_F012.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4219 + }, + { + "path": "Ses01F_script03_2_M022.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4220 + }, + { + "path": "Ses05F_script01_2_F014.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4221 + }, + { + "path": "Ses03M_script02_1_M035.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4222 + }, + { + "path": "Ses03M_script03_2_M011.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4223 + }, + { + "path": "Ses02M_impro08_M009.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4224 + }, + { + "path": "Ses05M_script01_1_F035.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4225 + }, + { + "path": "Ses03M_script01_3_F017.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4226 + }, + { + "path": "Ses03F_script02_1_M009.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4227 + }, + { + "path": "Ses02F_script02_2_M038.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4228 + }, + { + "path": "Ses05M_impro03_M029.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4229 + }, + { + "path": "Ses04F_impro08_M023.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4230 + }, + { + "path": "Ses04F_impro06_M009.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4231 + }, + { + "path": "Ses05F_script02_2_F032.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4232 + }, + { + "path": "Ses04M_impro06_M010.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4233 + }, + { + "path": "Ses01M_script02_2_M003.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4234 + }, + { + "path": "Ses05F_impro02_F010.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4235 + }, + { + "path": "Ses03M_impro06_F013.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4236 + }, + { + "path": "Ses04M_script01_3_M031.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4237 + }, + { + "path": "Ses03M_script03_2_F041.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4238 + }, + { + "path": "Ses03F_script01_3_M033.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4239 + }, + { + "path": "Ses05M_script02_2_M000.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4240 + }, + { + "path": "Ses03M_script03_2_F031.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4241 + }, + { + "path": "Ses05M_impro04_F012.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4242 + }, + { + "path": "Ses03F_impro03_F015.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4243 + }, + { + "path": "Ses02F_impro06_F002.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4244 + }, + { + "path": "Ses04M_impro01_F021.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4245 + }, + { + "path": "Ses01M_script02_1_M040.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4246 + }, + { + "path": "Ses01F_impro05_M017.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4247 + }, + { + "path": "Ses05M_impro06_M005.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4248 + }, + { + "path": "Ses03F_impro06_F021.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4249 + }, + { + "path": "Ses04F_script01_3_M047.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4250 + }, + { + "path": "Ses03M_script02_2_F016.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4251 + }, + { + "path": "Ses02F_script02_1_M002.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4252 + }, + { + "path": "Ses02F_impro04_M001.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4253 + }, + { + "path": "Ses01M_script03_1_M028.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4254 + }, + { + "path": "Ses05F_impro04_M017.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4255 + }, + { + "path": "Ses05F_script03_2_M019.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4256 + }, + { + "path": "Ses01M_impro04_F018.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4257 + }, + { + "path": "Ses01F_script02_1_M035.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4258 + }, + { + "path": "Ses01F_script02_1_M014.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4259 + }, + { + "path": "Ses02F_impro06_M008.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4260 + }, + { + "path": "Ses02F_impro07_F008.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4261 + }, + { + "path": "Ses05F_impro06_M001.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4262 + }, + { + "path": "Ses01F_impro05_M009.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4263 + }, + { + "path": "Ses03F_script01_3_M006.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4264 + }, + { + "path": "Ses05F_impro06_F019.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4265 + }, + { + "path": "Ses02F_impro05_M010.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4266 + }, + { + "path": "Ses04F_script03_1_F005.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4267 + }, + { + "path": "Ses02M_impro06_F004.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4268 + }, + { + "path": "Ses05F_impro03_F051.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4269 + }, + { + "path": "Ses05M_impro08_M008.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4270 + }, + { + "path": "Ses04F_script01_1_M022.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4271 + }, + { + "path": "Ses05M_script01_1b_M036.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4272 + }, + { + "path": "Ses04M_script03_2_F037.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4273 + }, + { + "path": "Ses02M_script03_2_M011.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4274 + }, + { + "path": "Ses01F_impro02_F020.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4275 + }, + { + "path": "Ses02F_impro03_M019.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4276 + }, + { + "path": "Ses03M_impro05a_M018.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4277 + }, + { + "path": "Ses04F_script01_1_M013.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4278 + }, + { + "path": "Ses05M_script01_1b_M003.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4279 + }, + { + "path": "Ses03M_script02_1_M000.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4280 + }, + { + "path": "Ses05F_script03_1_M022.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4281 + }, + { + "path": "Ses03F_impro04_M018.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4282 + }, + { + "path": "Ses01F_script01_1_F015.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4283 + }, + { + "path": "Ses04M_impro06_F006.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4284 + }, + { + "path": "Ses02F_impro08_F024.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4285 + }, + { + "path": "Ses02M_impro01_M018.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4286 + }, + { + "path": "Ses01M_impro06_F010.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4287 + }, + { + "path": "Ses05F_impro01_M003.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4288 + }, + { + "path": "Ses01F_impro04_M014.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4289 + }, + { + "path": "Ses03M_impro02_F024.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4290 + }, + { + "path": "Ses04M_impro03_M020.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4291 + }, + { + "path": "Ses02F_impro06_F007.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4292 + }, + { + "path": "Ses04F_impro01_F016.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4293 + }, + { + "path": "Ses01F_script02_2_F045.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4294 + }, + { + "path": "Ses02M_impro04_F010.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4295 + }, + { + "path": "Ses03M_impro03_F036.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4296 + }, + { + "path": "Ses01F_script03_1_M036.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4297 + }, + { + "path": "Ses05F_script02_2_F017.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4298 + }, + { + "path": "Ses01M_script02_2_M021.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4299 + }, + { + "path": "Ses02F_script03_1_M013.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4300 + }, + { + "path": "Ses05F_impro02_M033.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4301 + }, + { + "path": "Ses02M_impro03_M021.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4302 + }, + { + "path": "Ses01M_script02_2_F015.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4303 + }, + { + "path": "Ses02M_impro08_M028.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4304 + }, + { + "path": "Ses04M_script01_1_F040.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4305 + }, + { + "path": "Ses04F_script02_2_F043.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4306 + }, + { + "path": "Ses03M_script02_2_F009.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4307 + }, + { + "path": "Ses01F_impro07_F012.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4308 + }, + { + "path": "Ses05F_script01_3_F015.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4309 + }, + { + "path": "Ses03F_impro08_M006.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4310 + }, + { + "path": "Ses05M_impro07_M011.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4311 + }, + { + "path": "Ses01F_script01_3_M008.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4312 + }, + { + "path": "Ses03M_script01_1_M004.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4313 + }, + { + "path": "Ses04M_impro07_M042.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4314 + }, + { + "path": "Ses05F_impro04_M026.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4315 + }, + { + "path": "Ses01F_impro06_F010.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4316 + }, + { + "path": "Ses01M_impro07_F027.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4317 + }, + { + "path": "Ses03M_impro03_M039.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4318 + }, + { + "path": "Ses01M_impro03_M006.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4319 + }, + { + "path": "Ses04F_impro04_M009.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4320 + }, + { + "path": "Ses01M_impro06_M030.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4321 + }, + { + "path": "Ses03M_script02_2_F049.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4322 + }, + { + "path": "Ses05F_impro04_M000.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4323 + }, + { + "path": "Ses02F_script01_3_M028.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4324 + }, + { + "path": "Ses02M_script03_1_F028.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4325 + }, + { + "path": "Ses05M_impro03_M031.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4326 + }, + { + "path": "Ses01M_script03_2_F040.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4327 + }, + { + "path": "Ses04M_script01_3_M033.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4328 + }, + { + "path": "Ses02F_impro02_F010.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4329 + }, + { + "path": "Ses01M_impro06_M017.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4330 + }, + { + "path": "Ses03F_script02_2_M038.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4331 + }, + { + "path": "Ses02F_impro05_M011.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4332 + }, + { + "path": "Ses01M_impro02_M016.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4333 + }, + { + "path": "Ses01F_impro02_F004.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4334 + }, + { + "path": "Ses01M_impro06_F019.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4335 + }, + { + "path": "Ses03M_script01_1_M007.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4336 + }, + { + "path": "Ses04M_impro03_F006.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4337 + }, + { + "path": "Ses04F_script01_3_M024.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4338 + }, + { + "path": "Ses02F_script02_2_M041.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4339 + }, + { + "path": "Ses02M_script02_1_M015.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4340 + }, + { + "path": "Ses02M_impro04_M014.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4341 + }, + { + "path": "Ses02F_impro06_M015.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4342 + }, + { + "path": "Ses05M_script01_1b_M032.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4343 + }, + { + "path": "Ses05F_script03_1_F031.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4344 + }, + { + "path": "Ses01M_impro06_M015.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4345 + }, + { + "path": "Ses05M_script02_1_M033.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4346 + }, + { + "path": "Ses02M_impro06_M015.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4347 + }, + { + "path": "Ses02F_impro01_F018.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4348 + }, + { + "path": "Ses05F_impro06_F002.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4349 + }, + { + "path": "Ses05M_impro07_M028.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4350 + }, + { + "path": "Ses01M_script02_2_F031.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4351 + }, + { + "path": "Ses02F_script02_2_M029.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4352 + }, + { + "path": "Ses03M_impro03_M015.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4353 + }, + { + "path": "Ses04M_impro03_M006.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4354 + }, + { + "path": "Ses01M_script01_1_F040.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4355 + }, + { + "path": "Ses04M_impro01_F019.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4356 + }, + { + "path": "Ses03M_impro06_M015.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4357 + }, + { + "path": "Ses03M_script02_1_M003.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4358 + }, + { + "path": "Ses04F_script02_1_M021.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4359 + }, + { + "path": "Ses02F_impro03_M025.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4360 + }, + { + "path": "Ses01F_impro07_M018.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4361 + }, + { + "path": "Ses03M_impro02_F026.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4362 + }, + { + "path": "Ses02F_impro03_M020.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4363 + }, + { + "path": "Ses03F_impro02_F036.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4364 + }, + { + "path": "Ses04F_impro04_M008.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4365 + }, + { + "path": "Ses01M_impro07_M018.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4366 + }, + { + "path": "Ses02M_script03_1_M021.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4367 + }, + { + "path": "Ses02F_script02_2_F019.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4368 + }, + { + "path": "Ses05M_script03_1_F026.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4369 + }, + { + "path": "Ses05M_script01_1_F025.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4370 + }, + { + "path": "Ses01M_impro02_M007.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4371 + }, + { + "path": "Ses03F_script01_3_M036.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4372 + }, + { + "path": "Ses03M_script02_1_M041.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4373 + }, + { + "path": "Ses05M_script02_2_M034.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4374 + }, + { + "path": "Ses02F_script02_1_M021.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4375 + }, + { + "path": "Ses04F_script01_3_M011.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4376 + }, + { + "path": "Ses05M_script02_2_F012.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4377 + }, + { + "path": "Ses03F_script01_1_M028.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4378 + }, + { + "path": "Ses05F_impro02_F035.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4379 + }, + { + "path": "Ses05F_script02_1_M000.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4380 + }, + { + "path": "Ses03F_script02_1_M016.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4381 + }, + { + "path": "Ses05F_script01_3_M032.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4382 + }, + { + "path": "Ses03M_impro06_M016.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4383 + }, + { + "path": "Ses03F_impro06_F010.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4384 + }, + { + "path": "Ses01F_script03_2_F004.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4385 + }, + { + "path": "Ses05M_impro01_F004.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4386 + }, + { + "path": "Ses01M_impro04_F006.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4387 + }, + { + "path": "Ses04F_script01_3_M040.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4388 + }, + { + "path": "Ses02F_impro05_F006.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4389 + }, + { + "path": "Ses04M_script01_1_M014.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4390 + }, + { + "path": "Ses05M_script01_3_F013.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4391 + }, + { + "path": "Ses04M_script02_1_M022.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4392 + }, + { + "path": "Ses02F_impro07_F007.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4393 + }, + { + "path": "Ses01M_script02_1_F005.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4394 + }, + { + "path": "Ses01M_script02_2_F032.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4395 + }, + { + "path": "Ses01F_script02_2_F015.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4396 + }, + { + "path": "Ses02M_impro05_F017.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4397 + }, + { + "path": "Ses02F_impro02_F007.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4398 + }, + { + "path": "Ses05M_script01_1_F001.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4399 + }, + { + "path": "Ses01F_script03_1_M008.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4400 + }, + { + "path": "Ses01F_script03_2_M033.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4401 + }, + { + "path": "Ses05F_script01_3_M016.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4402 + }, + { + "path": "Ses02M_impro02_M000.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4403 + }, + { + "path": "Ses04M_impro03_M021.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4404 + }, + { + "path": "Ses05M_script02_1_M006.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4405 + }, + { + "path": "Ses05M_impro08_M012.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4406 + }, + { + "path": "Ses05M_script02_1_M001.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4407 + }, + { + "path": "Ses01F_impro05_F021.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4408 + }, + { + "path": "Ses02F_script02_2_M020.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4409 + }, + { + "path": "Ses04F_impro07_M080.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4410 + }, + { + "path": "Ses03F_impro02_F019.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4411 + }, + { + "path": "Ses01F_impro06_F025.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4412 + }, + { + "path": "Ses03F_script01_1_M022.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4413 + }, + { + "path": "Ses04F_impro01_M014.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4414 + }, + { + "path": "Ses04F_script02_2_M020.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4415 + }, + { + "path": "Ses03F_impro02_M001.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4416 + }, + { + "path": "Ses03M_impro05a_M026.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4417 + }, + { + "path": "Ses02M_script01_1_M024.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4418 + }, + { + "path": "Ses04M_script02_2_F007.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4419 + }, + { + "path": "Ses03M_script02_1_M038.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4420 + }, + { + "path": "Ses01M_impro02_F013.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4421 + }, + { + "path": "Ses04M_script03_2_M038.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4422 + }, + { + "path": "Ses05F_script01_3_M028.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4423 + }, + { + "path": "Ses01F_impro02_F002.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4424 + }, + { + "path": "Ses05F_impro03_F029.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4425 + }, + { + "path": "Ses05F_script01_3_F017.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4426 + }, + { + "path": "Ses02F_impro08_F002.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4427 + }, + { + "path": "Ses02M_impro03_F024.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4428 + }, + { + "path": "Ses05M_impro07_F013.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4429 + }, + { + "path": "Ses01M_script03_1_M019.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4430 + }, + { + "path": "Ses01F_impro07_M019.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4431 + }, + { + "path": "Ses05M_script02_1_M007.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4432 + }, + { + "path": "Ses02F_impro04_M007.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4433 + }, + { + "path": "Ses05M_impro08_M009.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4434 + }, + { + "path": "Ses05M_impro04_F026.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4435 + }, + { + "path": "Ses02F_impro08_F022.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4436 + }, + { + "path": "Ses03F_impro06_F029.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4437 + }, + { + "path": "Ses05F_impro08_F002.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4438 + }, + { + "path": "Ses03F_script01_3_M027.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4439 + }, + { + "path": "Ses03M_impro02_M029.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4440 + }, + { + "path": "Ses02F_script01_1_M011.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4441 + }, + { + "path": "Ses03M_impro07_M013.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4442 + }, + { + "path": "Ses04M_impro03_M009.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4443 + }, + { + "path": "Ses03M_impro08a_M004.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4444 + }, + { + "path": "Ses03F_script02_1_M041.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4445 + }, + { + "path": "Ses05F_impro08_F019.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4446 + }, + { + "path": "Ses02M_script02_2_F019.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4447 + }, + { + "path": "Ses01M_script02_2_M030.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4448 + }, + { + "path": "Ses01M_script01_1_M002.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4449 + }, + { + "path": "Ses05M_script02_2_F029.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4450 + }, + { + "path": "Ses03M_script01_3_F020.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4451 + }, + { + "path": "Ses03M_script03_2_F043.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4452 + }, + { + "path": "Ses03F_script01_1_M045.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4453 + }, + { + "path": "Ses02F_impro03_M013.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4454 + }, + { + "path": "Ses04M_impro06_F004.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4455 + }, + { + "path": "Ses02F_impro02_F013.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4456 + }, + { + "path": "Ses05F_script02_1_M001.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4457 + }, + { + "path": "Ses02M_script03_1_F022.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4458 + }, + { + "path": "Ses05F_impro05_M014.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4459 + }, + { + "path": "Ses02F_impro02_F012.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4460 + }, + { + "path": "Ses04M_script01_3_M028.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4461 + }, + { + "path": "Ses05M_impro05_M010.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4462 + }, + { + "path": "Ses05M_impro04_F006.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4463 + }, + { + "path": "Ses04M_script02_2_M039.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4464 + }, + { + "path": "Ses05F_impro01_F017.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4465 + }, + { + "path": "Ses05F_script02_1_M032.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4466 + }, + { + "path": "Ses01M_impro04_F013.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4467 + }, + { + "path": "Ses03M_script02_2_F003.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4468 + }, + { + "path": "Ses03F_impro06_F031.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4469 + }, + { + "path": "Ses04M_impro06_M011.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4470 + }, + { + "path": "Ses03M_impro06_M019.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4471 + }, + { + "path": "Ses05F_impro07_M011.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4472 + }, + { + "path": "Ses05M_impro03_M009.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4473 + }, + { + "path": "Ses04F_impro07_F030.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4474 + }, + { + "path": "Ses03M_script01_3_F007.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4475 + }, + { + "path": "Ses03M_impro08b_M019.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4476 + }, + { + "path": "Ses03F_script03_2_F039.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4477 + }, + { + "path": "Ses02M_script03_2_F002.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4478 + }, + { + "path": "Ses04F_script03_2_M044.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4479 + }, + { + "path": "Ses04F_script03_1_F028.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4480 + }, + { + "path": "Ses04F_impro01_M024.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4481 + }, + { + "path": "Ses01M_impro04_F019.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4482 + }, + { + "path": "Ses01M_script02_1_M007.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4483 + }, + { + "path": "Ses02M_script02_2_F034.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4484 + }, + { + "path": "Ses03F_script01_1_F053.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4485 + }, + { + "path": "Ses04M_script02_1_M032.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4486 + }, + { + "path": "Ses02M_impro06_M004.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4487 + }, + { + "path": "Ses04F_impro07_F024.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4488 + }, + { + "path": "Ses02M_script02_1_M036.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4489 + }, + { + "path": "Ses03M_impro08a_M027.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4490 + }, + { + "path": "Ses03M_script03_2_M045.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4491 + }, + { + "path": "Ses02M_script01_1_M027.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4492 + }, + { + "path": "Ses01M_script01_3_M009.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4493 + }, + { + "path": "Ses05M_impro05_M009.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4494 + }, + { + "path": "Ses02F_impro01_M017.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4495 + }, + { + "path": "Ses03F_impro02_F021.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4496 + }, + { + "path": "Ses02M_script01_1_M037.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4497 + }, + { + "path": "Ses03M_impro07_F013.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4498 + }, + { + "path": "Ses05F_impro06_M008.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4499 + }, + { + "path": "Ses02M_impro05_M016.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4500 + }, + { + "path": "Ses03M_script01_3_M008.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4501 + }, + { + "path": "Ses04F_impro03_F013.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4502 + }, + { + "path": "Ses05F_impro05_F024.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4503 + }, + { + "path": "Ses05F_script02_1_M031.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4504 + }, + { + "path": "Ses03M_script02_2_F036.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4505 + }, + { + "path": "Ses01F_script01_1_M019.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4506 + }, + { + "path": "Ses01F_impro02_M010.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4507 + }, + { + "path": "Ses03F_impro06_F028.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4508 + }, + { + "path": "Ses05F_script02_2_F005.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4509 + }, + { + "path": "Ses05F_impro08_M004.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4510 + }, + { + "path": "Ses02M_impro06_M002.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4511 + }, + { + "path": "Ses04F_impro04_M002.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4512 + }, + { + "path": "Ses03F_script03_1_M007.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4513 + }, + { + "path": "Ses05F_impro07_M020.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4514 + }, + { + "path": "Ses05M_impro06_M014.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4515 + }, + { + "path": "Ses04F_impro07_M058.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4516 + }, + { + "path": "Ses02F_script03_1_F026.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4517 + }, + { + "path": "Ses04F_impro07_M038.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4518 + }, + { + "path": "Ses01F_impro05_M035.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4519 + }, + { + "path": "Ses05M_script01_3_F007.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4520 + }, + { + "path": "Ses02F_impro05_F001.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4521 + }, + { + "path": "Ses03M_script01_1_M031.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4522 + }, + { + "path": "Ses04F_script01_2_F016.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4523 + }, + { + "path": "Ses02F_impro07_F019.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4524 + }, + { + "path": "Ses04F_script01_1_M032.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4525 + }, + { + "path": "Ses02M_impro03_M009.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4526 + }, + { + "path": "Ses01F_impro06_F024.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4527 + }, + { + "path": "Ses02F_script03_2_F039.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4528 + }, + { + "path": "Ses01F_impro07_M002.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4529 + }, + { + "path": "Ses04M_impro07_M003.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4530 + }, + { + "path": "Ses05F_script01_3_M033.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4531 + }, + { + "path": "Ses02M_script02_2_M016.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4532 + }, + { + "path": "Ses04M_impro03_M028.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4533 + }, + { + "path": "Ses01F_script02_2_F006.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4534 + }, + { + "path": "Ses01M_impro03_M021.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4535 + }, + { + "path": "Ses02M_impro03_F028.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4536 + }, + { + "path": "Ses01M_impro07_F006.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4537 + }, + { + "path": "Ses03F_impro08_F003.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4538 + }, + { + "path": "Ses01F_impro02_M005.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4539 + }, + { + "path": "Ses03F_impro02_F029.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4540 + }, + { + "path": "Ses05F_impro07_M009.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4541 + }, + { + "path": "Ses02M_script03_2_F007.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4542 + }, + { + "path": "Ses01F_impro03_F021.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4543 + }, + { + "path": "Ses05M_impro05_F014.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4544 + }, + { + "path": "Ses03M_impro08b_M010.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4545 + }, + { + "path": "Ses01M_script01_3_F006.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4546 + }, + { + "path": "Ses02F_script02_2_M019.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4547 + }, + { + "path": "Ses05M_script01_1b_M005.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4548 + }, + { + "path": "Ses02M_impro08_M020.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4549 + }, + { + "path": "Ses01F_script03_1_M010.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4550 + }, + { + "path": "Ses03M_impro03_M020.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4551 + }, + { + "path": "Ses02F_impro08_F023.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4552 + }, + { + "path": "Ses02F_impro05_M002.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4553 + }, + { + "path": "Ses02M_impro04_F005.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4554 + }, + { + "path": "Ses05M_impro08_M017.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4555 + }, + { + "path": "Ses04F_script01_2_F010.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4556 + }, + { + "path": "Ses04M_impro07_M036.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4557 + }, + { + "path": "Ses03F_script01_3_M013.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4558 + }, + { + "path": "Ses05M_impro03_F007.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4559 + }, + { + "path": "Ses05F_impro01_F020.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4560 + }, + { + "path": "Ses02F_impro02_F002.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4561 + }, + { + "path": "Ses05M_impro07_M009.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4562 + }, + { + "path": "Ses02F_script01_2_F012.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4563 + }, + { + "path": "Ses01M_impro04_M023.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4564 + }, + { + "path": "Ses04M_impro08_M017.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4565 + }, + { + "path": "Ses05M_script01_1_M013.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4566 + }, + { + "path": "Ses02F_script01_3_M027.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4567 + }, + { + "path": "Ses04F_impro04_M020.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4568 + }, + { + "path": "Ses03M_impro06_M003.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4569 + }, + { + "path": "Ses03M_script02_2_F041.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4570 + }, + { + "path": "Ses04F_script01_2_M017.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4571 + }, + { + "path": "Ses01M_script02_1_M028.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4572 + }, + { + "path": "Ses03F_script01_3_F023.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4573 + }, + { + "path": "Ses05F_script02_1_M030.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4574 + }, + { + "path": "Ses01F_impro07_M007.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4575 + }, + { + "path": "Ses04F_impro07_M009.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4576 + }, + { + "path": "Ses05M_impro02_M030.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4577 + }, + { + "path": "Ses03M_impro06_M010.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4578 + }, + { + "path": "Ses02M_script03_2_F037.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4579 + }, + { + "path": "Ses03M_impro07_F011.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4580 + }, + { + "path": "Ses05F_impro06_M025.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4581 + }, + { + "path": "Ses02F_impro05_M021.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4582 + }, + { + "path": "Ses01M_script02_1_M001.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4583 + }, + { + "path": "Ses05M_impro07_M019.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4584 + }, + { + "path": "Ses01M_impro03_F008.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4585 + }, + { + "path": "Ses05M_impro08_M014.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4586 + }, + { + "path": "Ses03M_script02_1_M005.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4587 + }, + { + "path": "Ses03F_impro05_M013.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4588 + }, + { + "path": "Ses05F_impro04_M005.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4589 + }, + { + "path": "Ses05F_impro04_M025.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4590 + }, + { + "path": "Ses02F_script01_1_M038.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4591 + }, + { + "path": "Ses05M_impro03_M016.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4592 + }, + { + "path": "Ses05M_script02_2_F010.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4593 + }, + { + "path": "Ses03F_impro02_M029.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4594 + }, + { + "path": "Ses05M_script01_3_F016.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4595 + }, + { + "path": "Ses02M_script03_1_M019.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4596 + }, + { + "path": "Ses03F_impro08_M008.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4597 + }, + { + "path": "Ses03F_script01_2_M008.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4598 + }, + { + "path": "Ses03F_impro06_F002.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4599 + }, + { + "path": "Ses01M_script02_1_M041.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4600 + }, + { + "path": "Ses04M_impro08_F023.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4601 + }, + { + "path": "Ses03M_impro08a_M012.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4602 + }, + { + "path": "Ses04F_impro08_F004.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4603 + }, + { + "path": "Ses01M_impro07_M000.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4604 + }, + { + "path": "Ses01F_script02_1_M011.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4605 + }, + { + "path": "Ses02F_impro04_M005.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4606 + }, + { + "path": "Ses04F_impro07_M077.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4607 + }, + { + "path": "Ses03F_impro07_F013.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4608 + }, + { + "path": "Ses02F_script02_2_F034.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4609 + }, + { + "path": "Ses02F_impro03_F014.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4610 + }, + { + "path": "Ses01F_script01_2_M017.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4611 + }, + { + "path": "Ses04M_script02_2_M040.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4612 + }, + { + "path": "Ses03F_impro04_M015.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4613 + }, + { + "path": "Ses02M_script01_3_M022.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4614 + }, + { + "path": "Ses04M_script01_3_M034.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4615 + }, + { + "path": "Ses04F_impro05_F019.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4616 + }, + { + "path": "Ses01F_script01_3_F001.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4617 + }, + { + "path": "Ses05F_script02_2_M031.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4618 + }, + { + "path": "Ses04M_impro03_F011.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4619 + }, + { + "path": "Ses02M_script02_2_F013.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4620 + }, + { + "path": "Ses03M_impro05b_M026.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4621 + }, + { + "path": "Ses03F_impro02_F007.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4622 + }, + { + "path": "Ses01M_script01_1_F014.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4623 + }, + { + "path": "Ses02M_impro06_F012.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4624 + }, + { + "path": "Ses01F_script01_1_M024.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4625 + }, + { + "path": "Ses04F_impro02_M026.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4626 + }, + { + "path": "Ses01F_impro01_M013.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4627 + }, + { + "path": "Ses02M_impro05_M002.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4628 + }, + { + "path": "Ses04F_impro04_M022.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4629 + }, + { + "path": "Ses02F_script01_3_M010.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4630 + }, + { + "path": "Ses03M_impro03_M018.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4631 + }, + { + "path": "Ses02F_impro06_M009.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4632 + }, + { + "path": "Ses01M_script03_2_F023.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4633 + }, + { + "path": "Ses05F_impro05_M008.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4634 + }, + { + "path": "Ses04F_impro04_M021.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4635 + }, + { + "path": "Ses05M_impro03_M011.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4636 + }, + { + "path": "Ses01M_script01_2_F005.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4637 + }, + { + "path": "Ses02M_script02_1_M024.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4638 + }, + { + "path": "Ses02F_script01_1_F027.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4639 + }, + { + "path": "Ses05F_impro03_M010.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4640 + }, + { + "path": "Ses03M_impro08b_M013.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4641 + }, + { + "path": "Ses04F_script01_1_M010.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4642 + }, + { + "path": "Ses03M_impro03_M031.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4643 + }, + { + "path": "Ses02F_impro01_M001.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4644 + }, + { + "path": "Ses05M_script02_2_M035.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4645 + }, + { + "path": "Ses04M_script02_1_M021.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4646 + }, + { + "path": "Ses02M_impro06_M016.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4647 + }, + { + "path": "Ses01M_impro05_M025.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4648 + }, + { + "path": "Ses01M_script03_1_F037.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4649 + }, + { + "path": "Ses05F_impro05_M052.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4650 + }, + { + "path": "Ses03M_impro05b_M011.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4651 + }, + { + "path": "Ses04M_impro06_F007.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4652 + }, + { + "path": "Ses01M_impro03_F010.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4653 + }, + { + "path": "Ses05M_impro03_M030.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4654 + }, + { + "path": "Ses04F_impro07_M071.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4655 + }, + { + "path": "Ses04F_script01_3_F010.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4656 + }, + { + "path": "Ses04F_impro03_M042.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4657 + }, + { + "path": "Ses03F_script02_1_M021.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4658 + }, + { + "path": "Ses04F_script02_2_F046.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4659 + }, + { + "path": "Ses01F_script03_2_M020.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4660 + }, + { + "path": "Ses03M_script03_2_F045.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4661 + }, + { + "path": "Ses05F_impro03_M037.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4662 + }, + { + "path": "Ses02M_impro03_F008.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4663 + }, + { + "path": "Ses02M_script01_3_M027.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4664 + }, + { + "path": "Ses03M_script02_2_F038.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4665 + }, + { + "path": "Ses02F_impro07_M005.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4666 + }, + { + "path": "Ses03F_script02_2_F016.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4667 + }, + { + "path": "Ses03M_impro02_F022.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4668 + }, + { + "path": "Ses05F_impro07_F028.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4669 + }, + { + "path": "Ses03M_impro05a_M016.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4670 + }, + { + "path": "Ses01M_script02_2_M038.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4671 + }, + { + "path": "Ses01M_impro01_M032.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4672 + }, + { + "path": "Ses01F_script03_1_M007.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4673 + }, + { + "path": "Ses05F_script01_1_F007.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4674 + }, + { + "path": "Ses01M_script03_1_M037.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4675 + }, + { + "path": "Ses04F_script03_2_F027.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4676 + }, + { + "path": "Ses03M_impro02_F030.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4677 + }, + { + "path": "Ses03M_impro08b_M012.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4678 + }, + { + "path": "Ses01M_script01_1_M020.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4679 + }, + { + "path": "Ses05F_script02_2_F016.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4680 + }, + { + "path": "Ses03M_impro08a_M000.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4681 + }, + { + "path": "Ses05F_impro08_F022.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4682 + }, + { + "path": "Ses05M_impro02_M012.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4683 + }, + { + "path": "Ses05M_impro07_M043.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4684 + }, + { + "path": "Ses05M_script02_1_M020.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4685 + }, + { + "path": "Ses02M_script03_2_F015.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4686 + }, + { + "path": "Ses02F_impro01_F000.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4687 + }, + { + "path": "Ses05M_impro06_M018.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4688 + }, + { + "path": "Ses02F_impro03_M024.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4689 + }, + { + "path": "Ses02M_impro07_M025.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4690 + }, + { + "path": "Ses01M_impro07_M005.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4691 + }, + { + "path": "Ses03F_script02_1_M002.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4692 + }, + { + "path": "Ses01F_script03_2_F024.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4693 + }, + { + "path": "Ses04M_script02_1_F004.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4694 + }, + { + "path": "Ses02F_script02_1_M001.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4695 + }, + { + "path": "Ses04F_impro04_M003.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4696 + }, + { + "path": "Ses04F_impro01_F012.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4697 + }, + { + "path": "Ses05M_script03_1_M021.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4698 + }, + { + "path": "Ses03M_script01_1_M038.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4699 + }, + { + "path": "Ses02M_impro04_F007.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4700 + }, + { + "path": "Ses01F_script01_1_M007.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4701 + }, + { + "path": "Ses05F_impro04_M001.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4702 + }, + { + "path": "Ses03M_script01_1_M009.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4703 + }, + { + "path": "Ses02F_impro04_M010.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4704 + }, + { + "path": "Ses04M_script03_2_M034.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4705 + }, + { + "path": "Ses05M_impro08_M010.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4706 + }, + { + "path": "Ses02M_impro06_M013.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4707 + }, + { + "path": "Ses03M_impro05b_M031.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4708 + }, + { + "path": "Ses02M_script02_2_M020.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4709 + }, + { + "path": "Ses03M_script03_2_F042.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4710 + }, + { + "path": "Ses03F_script03_1_F026.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4711 + }, + { + "path": "Ses04M_impro08_M026.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4712 + }, + { + "path": "Ses05M_impro05_F003.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4713 + }, + { + "path": "Ses05F_impro07_F008.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4714 + }, + { + "path": "Ses01F_impro06_F009.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4715 + }, + { + "path": "Ses05M_script03_2_M020.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4716 + }, + { + "path": "Ses05M_impro03_F021.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4717 + }, + { + "path": "Ses05F_impro06_F014.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4718 + }, + { + "path": "Ses04F_impro02_F007.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4719 + }, + { + "path": "Ses03F_impro06_F020.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4720 + }, + { + "path": "Ses03M_script02_2_F047.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4721 + }, + { + "path": "Ses02F_script03_1_M023.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4722 + }, + { + "path": "Ses01M_script01_1_F023.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4723 + }, + { + "path": "Ses03F_script03_1_M022.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4724 + }, + { + "path": "Ses02M_script02_1_M035.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4725 + }, + { + "path": "Ses01F_impro05_F017.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4726 + }, + { + "path": "Ses02M_impro07_M031.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4727 + }, + { + "path": "Ses01M_impro02_M023.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4728 + }, + { + "path": "Ses01F_script03_2_M025.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4729 + }, + { + "path": "Ses04M_impro06_F019.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4730 + }, + { + "path": "Ses04M_script02_2_F016.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4731 + }, + { + "path": "Ses03F_impro06_F015.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4732 + }, + { + "path": "Ses05M_script01_1_M034.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4733 + }, + { + "path": "Ses04F_impro03_M039.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4734 + }, + { + "path": "Ses01M_impro03_F023.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4735 + }, + { + "path": "Ses03F_script01_3_M038.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4736 + }, + { + "path": "Ses03F_impro07_F007.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4737 + }, + { + "path": "Ses03F_impro07_M019.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4738 + }, + { + "path": "Ses05M_script02_1_M023.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4739 + }, + { + "path": "Ses05F_script02_1_M019.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4740 + }, + { + "path": "Ses04F_script02_1_M036.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4741 + }, + { + "path": "Ses03F_script01_3_M035.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4742 + }, + { + "path": "Ses05M_impro02_M025.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4743 + }, + { + "path": "Ses02M_impro08_M012.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4744 + }, + { + "path": "Ses02F_impro07_F011.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4745 + }, + { + "path": "Ses02F_script02_1_M015.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4746 + }, + { + "path": "Ses02M_script01_3_M020.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4747 + }, + { + "path": "Ses01M_script02_1_M031.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4748 + }, + { + "path": "Ses03F_impro07_F016.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4749 + }, + { + "path": "Ses05F_impro08_F014.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4750 + }, + { + "path": "Ses02F_impro07_F018.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4751 + }, + { + "path": "Ses01M_impro05_M026.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4752 + }, + { + "path": "Ses04F_impro02_M024.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4753 + }, + { + "path": "Ses05F_script01_3_M005.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4754 + }, + { + "path": "Ses01M_script02_2_F044.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4755 + }, + { + "path": "Ses02M_impro06_M008.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4756 + }, + { + "path": "Ses03F_script03_2_F017.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4757 + }, + { + "path": "Ses05F_script01_3_F008.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4758 + }, + { + "path": "Ses02M_impro06_M007.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4759 + }, + { + "path": "Ses01M_impro07_M003.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4760 + }, + { + "path": "Ses03M_script03_2_M031.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4761 + }, + { + "path": "Ses05F_script02_2_M005.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4762 + }, + { + "path": "Ses03M_impro08a_M002.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4763 + }, + { + "path": "Ses02F_script02_1_M028.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4764 + }, + { + "path": "Ses03M_script02_1_M016.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4765 + }, + { + "path": "Ses03M_script03_2_M039.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4766 + }, + { + "path": "Ses03M_script01_3_M033.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4767 + }, + { + "path": "Ses03M_script02_1_M002.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4768 + }, + { + "path": "Ses05M_impro07_F012.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4769 + }, + { + "path": "Ses01M_impro02_M002.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4770 + }, + { + "path": "Ses04F_script01_1_F024.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4771 + }, + { + "path": "Ses05F_impro08_F008.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4772 + }, + { + "path": "Ses03M_impro06_M008.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4773 + }, + { + "path": "Ses03M_impro02_F033.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4774 + }, + { + "path": "Ses02F_script02_2_F005.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4775 + }, + { + "path": "Ses03M_script02_2_F033.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4776 + }, + { + "path": "Ses01F_script03_2_M036.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4777 + }, + { + "path": "Ses01M_script01_1_F030.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4778 + }, + { + "path": "Ses03F_script01_1_M002.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4779 + }, + { + "path": "Ses01M_script01_3_M024.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4780 + }, + { + "path": "Ses03M_impro05b_M023.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4781 + }, + { + "path": "Ses02M_impro07_M032.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4782 + }, + { + "path": "Ses04M_impro06_F002.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4783 + }, + { + "path": "Ses05F_impro03_M027.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4784 + }, + { + "path": "Ses05F_impro06_M022.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4785 + }, + { + "path": "Ses03M_impro01_M023.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4786 + }, + { + "path": "Ses01F_impro03_M009.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4787 + }, + { + "path": "Ses01F_script02_1_M016.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4788 + }, + { + "path": "Ses03M_script01_1_M036.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4789 + }, + { + "path": "Ses04F_script01_3_M009.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4790 + }, + { + "path": "Ses03M_impro02_F028.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4791 + }, + { + "path": "Ses02M_script03_2_F041.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4792 + }, + { + "path": "Ses05F_script01_1_M003.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4793 + }, + { + "path": "Ses02F_script02_2_F006.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4794 + }, + { + "path": "Ses05F_script01_3_F002.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4795 + }, + { + "path": "Ses04F_script01_3_M041.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4796 + }, + { + "path": "Ses04M_script03_1_M021.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4797 + }, + { + "path": "Ses03M_script01_3_M040.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4798 + }, + { + "path": "Ses02F_impro08_M016.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4799 + }, + { + "path": "Ses02M_impro07_M020.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4800 + }, + { + "path": "Ses02M_script01_3_F019.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4801 + }, + { + "path": "Ses05M_impro07_M045.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4802 + }, + { + "path": "Ses03F_impro08_F014.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4803 + }, + { + "path": "Ses02M_script01_3_M032.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4804 + }, + { + "path": "Ses05F_impro03_M013.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4805 + }, + { + "path": "Ses03M_script01_1_F030.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4806 + }, + { + "path": "Ses04F_script01_3_M039.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4807 + }, + { + "path": "Ses05F_script02_2_M036.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4808 + }, + { + "path": "Ses02M_script02_2_F016.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4809 + }, + { + "path": "Ses04F_impro06_M006.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4810 + }, + { + "path": "Ses01M_impro02_F017.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4811 + }, + { + "path": "Ses01M_script03_2_F041.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4812 + }, + { + "path": "Ses04F_script03_1_M008.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4813 + }, + { + "path": "Ses02M_script02_1_M005.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4814 + }, + { + "path": "Ses05M_impro04_F009.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4815 + }, + { + "path": "Ses03M_impro02_M015.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4816 + }, + { + "path": "Ses03F_impro08_F028.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4817 + }, + { + "path": "Ses03M_script02_2_M017.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4818 + }, + { + "path": "Ses03F_impro01_M002.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4819 + }, + { + "path": "Ses05F_impro05_F018.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4820 + }, + { + "path": "Ses01M_impro04_F010.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4821 + }, + { + "path": "Ses03M_script02_1_M024.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4822 + }, + { + "path": "Ses02F_script03_2_F038.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4823 + }, + { + "path": "Ses05F_impro05_M023.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4824 + }, + { + "path": "Ses04M_script01_2_F011.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4825 + }, + { + "path": "Ses03F_impro04_M010.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4826 + }, + { + "path": "Ses02M_impro05_F012.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4827 + }, + { + "path": "Ses03F_script03_2_F042.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4828 + }, + { + "path": "Ses05F_impro06_M002.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4829 + }, + { + "path": "Ses04F_impro04_M018.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4830 + }, + { + "path": "Ses04F_script01_3_M043.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4831 + }, + { + "path": "Ses01M_script02_2_M023.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4832 + }, + { + "path": "Ses03F_script02_2_F019.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4833 + }, + { + "path": "Ses01F_script01_1_M002.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4834 + }, + { + "path": "Ses05M_impro04_F008.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4835 + }, + { + "path": "Ses04F_script01_3_M031.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4836 + }, + { + "path": "Ses02M_impro07_F001.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4837 + }, + { + "path": "Ses01M_impro03_M023.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4838 + }, + { + "path": "Ses03M_script01_3_M023.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4839 + }, + { + "path": "Ses05F_impro08_F023.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4840 + }, + { + "path": "Ses01M_script02_1_M032.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4841 + }, + { + "path": "Ses03F_script01_1_M018.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4842 + }, + { + "path": "Ses03M_impro08a_F020.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4843 + }, + { + "path": "Ses01M_impro04_M018.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4844 + }, + { + "path": "Ses05M_script01_3_M017.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4845 + }, + { + "path": "Ses04M_script02_2_F019.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4846 + }, + { + "path": "Ses03F_impro06_M011.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4847 + }, + { + "path": "Ses02M_impro06_M017.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4848 + }, + { + "path": "Ses05F_impro07_F012.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4849 + }, + { + "path": "Ses03F_impro06_M002.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4850 + }, + { + "path": "Ses05M_impro05_M000.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4851 + }, + { + "path": "Ses02M_impro02_M001.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4852 + }, + { + "path": "Ses05F_script03_1_M029.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4853 + }, + { + "path": "Ses05M_impro03_M012.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4854 + }, + { + "path": "Ses02F_script01_2_M002.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4855 + }, + { + "path": "Ses02M_script02_2_F005.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4856 + }, + { + "path": "Ses03M_script01_1_F021.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4857 + }, + { + "path": "Ses03M_impro08a_M006.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4858 + }, + { + "path": "Ses01M_impro01_F022.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4859 + }, + { + "path": "Ses05F_script03_1_M005.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4860 + }, + { + "path": "Ses02F_impro03_F021.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4861 + }, + { + "path": "Ses04M_impro07_M012.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4862 + }, + { + "path": "Ses05M_impro04_M020.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4863 + }, + { + "path": "Ses03F_impro02_F031.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4864 + }, + { + "path": "Ses02F_impro08_F004.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4865 + }, + { + "path": "Ses05M_script01_3_F001.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4866 + }, + { + "path": "Ses01F_script01_2_F008.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4867 + }, + { + "path": "Ses02F_script02_2_F015.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4868 + }, + { + "path": "Ses04M_impro03_F003.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4869 + }, + { + "path": "Ses03M_impro07_F017.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4870 + }, + { + "path": "Ses04M_impro03_M030.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4871 + }, + { + "path": "Ses05F_impro04_M041.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4872 + }, + { + "path": "Ses05M_impro08_M006.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4873 + }, + { + "path": "Ses02M_script01_2_F001.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4874 + }, + { + "path": "Ses01M_script02_1_F022.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4875 + }, + { + "path": "Ses03M_impro05b_M014.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4876 + }, + { + "path": "Ses05M_script03_2_M041.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4877 + }, + { + "path": "Ses01F_impro06_F026.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4878 + }, + { + "path": "Ses03F_script01_3_M000.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4879 + }, + { + "path": "Ses05F_script01_1_M013.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4880 + }, + { + "path": "Ses02F_impro08_M005.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4881 + }, + { + "path": "Ses03F_script01_3_M014.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4882 + }, + { + "path": "Ses02M_script03_2_F039.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4883 + }, + { + "path": "Ses03M_impro03_F010.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4884 + }, + { + "path": "Ses05F_impro03_M036.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4885 + }, + { + "path": "Ses04M_impro06_M002.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4886 + }, + { + "path": "Ses05F_impro06_F011.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4887 + }, + { + "path": "Ses02F_impro07_F010.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4888 + }, + { + "path": "Ses01M_script01_3_F008.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4889 + }, + { + "path": "Ses01M_script03_2_M029.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4890 + }, + { + "path": "Ses04M_script03_2_F049.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4891 + }, + { + "path": "Ses01F_impro05_F019.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4892 + }, + { + "path": "Ses03M_script01_3_M034.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4893 + }, + { + "path": "Ses02F_script02_1_M006.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4894 + }, + { + "path": "Ses01F_impro07_F016.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4895 + }, + { + "path": "Ses03M_script01_2_F012.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4896 + }, + { + "path": "Ses01M_script02_1_M014.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4897 + }, + { + "path": "Ses03M_script01_1_F038.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4898 + }, + { + "path": "Ses05M_impro06_M007.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4899 + }, + { + "path": "Ses01F_script01_2_F012.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4900 + }, + { + "path": "Ses03F_script01_3_F002.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4901 + }, + { + "path": "Ses04F_script02_2_F005.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4902 + }, + { + "path": "Ses03M_script03_2_M042.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4903 + }, + { + "path": "Ses03F_script03_1_F019.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4904 + }, + { + "path": "Ses01M_impro05_F020.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4905 + }, + { + "path": "Ses02M_script02_1_M038.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4906 + }, + { + "path": "Ses05F_impro02_F029.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4907 + }, + { + "path": "Ses05F_impro04_F000.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4908 + }, + { + "path": "Ses03M_impro03_F018.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4909 + }, + { + "path": "Ses04F_script03_1_F023.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4910 + }, + { + "path": "Ses04M_impro04_F008.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4911 + }, + { + "path": "Ses02F_script01_3_F020.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4912 + }, + { + "path": "Ses02F_impro06_M017.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4913 + }, + { + "path": "Ses05F_impro02_F023.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4914 + }, + { + "path": "Ses04F_impro07_F018.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4915 + }, + { + "path": "Ses03F_impro05_F005.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4916 + }, + { + "path": "Ses02M_impro02_F009.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4917 + }, + { + "path": "Ses05F_script01_3_M040.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4918 + }, + { + "path": "Ses01M_impro06_M011.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4919 + }, + { + "path": "Ses03F_script02_1_M038.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4920 + }, + { + "path": "Ses01F_impro02_M001.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4921 + }, + { + "path": "Ses04M_impro06_F014.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4922 + }, + { + "path": "Ses04F_script01_3_M035.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4923 + }, + { + "path": "Ses05F_impro07_M032.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4924 + }, + { + "path": "Ses05F_script03_1_M028.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4925 + }, + { + "path": "Ses01F_script02_2_F017.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4926 + }, + { + "path": "Ses05M_script01_1_M033.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4927 + }, + { + "path": "Ses04F_script03_2_F022.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4928 + }, + { + "path": "Ses02M_impro07_M026.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4929 + }, + { + "path": "Ses05M_impro07_F020.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4930 + }, + { + "path": "Ses01F_script02_1_M032.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4931 + }, + { + "path": "Ses03F_script01_1_F052.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4932 + }, + { + "path": "Ses02F_impro05_M003.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4933 + }, + { + "path": "Ses04M_script02_1_M008.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4934 + }, + { + "path": "Ses01M_script01_3_M039.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4935 + }, + { + "path": "Ses02M_script02_2_M017.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4936 + }, + { + "path": "Ses03F_script02_1_M015.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4937 + }, + { + "path": "Ses04M_script01_1_M034.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4938 + }, + { + "path": "Ses05F_impro02_F038.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4939 + }, + { + "path": "Ses05F_impro06_F021.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4940 + }, + { + "path": "Ses03M_impro05b_M013.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4941 + }, + { + "path": "Ses05M_impro06_M011.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4942 + }, + { + "path": "Ses04M_script02_1_M029.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4943 + }, + { + "path": "Ses01F_impro07_M006.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4944 + }, + { + "path": "Ses04F_script02_2_F002.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4945 + }, + { + "path": "Ses02F_impro08_M015.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4946 + }, + { + "path": "Ses02F_impro06_F013.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4947 + }, + { + "path": "Ses03F_impro06_F003.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4948 + }, + { + "path": "Ses04M_script01_3_M027.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4949 + }, + { + "path": "Ses01M_script02_2_F036.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4950 + }, + { + "path": "Ses03M_impro06_M017.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4951 + }, + { + "path": "Ses05F_script02_2_M033.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4952 + }, + { + "path": "Ses04M_impro08_M027.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4953 + }, + { + "path": "Ses05F_impro07_F023.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4954 + }, + { + "path": "Ses04F_script03_1_F024.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4955 + }, + { + "path": "Ses02F_script02_1_M025.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4956 + }, + { + "path": "Ses01M_script02_2_M051.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4957 + }, + { + "path": "Ses05F_script02_2_F026.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4958 + }, + { + "path": "Ses01F_script01_3_M023.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4959 + }, + { + "path": "Ses05F_script01_3_M035.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4960 + }, + { + "path": "Ses02F_impro08_F010.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4961 + }, + { + "path": "Ses02M_script01_3_F026.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4962 + }, + { + "path": "Ses01M_impro05_F034.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4963 + }, + { + "path": "Ses05F_impro08_F017.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4964 + }, + { + "path": "Ses04F_script02_1_F014.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4965 + }, + { + "path": "Ses05F_impro01_F008.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4966 + }, + { + "path": "Ses04F_script02_2_F008.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4967 + }, + { + "path": "Ses03F_script02_1_M014.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4968 + }, + { + "path": "Ses04M_script01_3_M024.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4969 + }, + { + "path": "Ses03F_impro06_F013.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4970 + }, + { + "path": "Ses03M_impro01_M015.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4971 + }, + { + "path": "Ses05F_impro03_M012.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "iemocap", + "uniq_id": 4972 + }, + { + "path": "dia210_utt0.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4973 + }, + { + "path": "dia262_utt10.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4974 + }, + { + "path": "dia127_utt5.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4975 + }, + { + "path": "dia14_utt9.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4976 + }, + { + "path": "dia97_utt10.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4977 + }, + { + "path": "dia189_utt0.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4978 + }, + { + "path": "dia167_utt7.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4979 + }, + { + "path": "dia137_utt9.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4980 + }, + { + "path": "dia229_utt4.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4981 + }, + { + "path": "dia241_utt11.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4982 + }, + { + "path": "dia139_utt5.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4983 + }, + { + "path": "dia217_utt2.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4984 + }, + { + "path": "dia98_utt8.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4985 + }, + { + "path": "dia264_utt1.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4986 + }, + { + "path": "dia268_utt5.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4987 + }, + { + "path": "dia157_utt2.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4988 + }, + { + "path": "dia65_utt10.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4989 + }, + { + "path": "dia112_utt2.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4990 + }, + { + "path": "dia64_utt2.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4991 + }, + { + "path": "dia155_utt4.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4992 + }, + { + "path": "dia208_utt2.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4993 + }, + { + "path": "dia151_utt3.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4994 + }, + { + "path": "dia244_utt10.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4995 + }, + { + "path": "dia41_utt4.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4996 + }, + { + "path": "dia52_utt1.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4997 + }, + { + "path": "dia102_utt7.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4998 + }, + { + "path": "dia248_utt0.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 4999 + }, + { + "path": "dia196_utt0.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5000 + }, + { + "path": "dia105_utt0.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5001 + }, + { + "path": "dia156_utt3.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5002 + }, + { + "path": "dia229_utt1.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5003 + }, + { + "path": "dia206_utt0.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5004 + }, + { + "path": "dia173_utt15.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5005 + }, + { + "path": "dia155_utt2.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5006 + }, + { + "path": "dia150_utt1.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5007 + }, + { + "path": "dia25_utt4.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5008 + }, + { + "path": "dia139_utt7.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5009 + }, + { + "path": "dia261_utt4.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5010 + }, + { + "path": "dia237_utt5.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5011 + }, + { + "path": "dia75_utt2.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5012 + }, + { + "path": "dia253_utt10.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5013 + }, + { + "path": "dia69_utt3.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5014 + }, + { + "path": "dia168_utt1.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5015 + }, + { + "path": "dia222_utt8.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5016 + }, + { + "path": "dia252_utt1.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5017 + }, + { + "path": "dia278_utt3.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5018 + }, + { + "path": "dia253_utt12.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5019 + }, + { + "path": "dia7_utt7.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5020 + }, + { + "path": "dia191_utt0.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5021 + }, + { + "path": "dia201_utt4.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5022 + }, + { + "path": "dia29_utt12.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5023 + }, + { + "path": "dia100_utt14.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5024 + }, + { + "path": "dia168_utt13.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5025 + }, + { + "path": "dia138_utt2.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5026 + }, + { + "path": "dia143_utt4.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5027 + }, + { + "path": "dia119_utt3.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5028 + }, + { + "path": "dia136_utt0.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5029 + }, + { + "path": "dia185_utt5.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5030 + }, + { + "path": "dia85_utt3.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5031 + }, + { + "path": "dia56_utt1.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5032 + }, + { + "path": "dia73_utt2.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5033 + }, + { + "path": "dia166_utt11.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5034 + }, + { + "path": "dia167_utt5.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5035 + }, + { + "path": "dia242_utt3.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5036 + }, + { + "path": "dia247_utt3.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5037 + }, + { + "path": "dia130_utt17.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5038 + }, + { + "path": "dia7_utt8.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5039 + }, + { + "path": "dia277_utt0.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5040 + }, + { + "path": "dia132_utt16.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5041 + }, + { + "path": "dia107_utt1.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5042 + }, + { + "path": "dia71_utt9.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5043 + }, + { + "path": "dia154_utt5.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5044 + }, + { + "path": "dia186_utt1.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5045 + }, + { + "path": "dia279_utt7.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5046 + }, + { + "path": "dia244_utt12.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5047 + }, + { + "path": "dia260_utt4.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5048 + }, + { + "path": "dia271_utt5.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5049 + }, + { + "path": "dia163_utt4.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5050 + }, + { + "path": "dia132_utt19.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5051 + }, + { + "path": "dia71_utt11.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5052 + }, + { + "path": "dia223_utt5.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5053 + }, + { + "path": "dia109_utt6.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5054 + }, + { + "path": "dia100_utt0.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5055 + }, + { + "path": "dia123_utt1.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5056 + }, + { + "path": "dia96_utt16.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5057 + }, + { + "path": "dia196_utt17.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5058 + }, + { + "path": "dia160_utt12.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5059 + }, + { + "path": "dia110_utt7.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5060 + }, + { + "path": "dia137_utt3.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5061 + }, + { + "path": "dia77_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5062 + }, + { + "path": "dia166_utt6.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5063 + }, + { + "path": "dia196_utt13.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5064 + }, + { + "path": "dia125_utt21.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5065 + }, + { + "path": "dia65_utt8.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5066 + }, + { + "path": "dia76_utt12.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5067 + }, + { + "path": "dia233_utt7.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5068 + }, + { + "path": "dia108_utt6.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5069 + }, + { + "path": "dia17_utt10.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5070 + }, + { + "path": "dia195_utt2.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5071 + }, + { + "path": "dia173_utt4.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5072 + }, + { + "path": "dia52_utt11.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5073 + }, + { + "path": "dia270_utt6.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5074 + }, + { + "path": "dia68_utt6.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5075 + }, + { + "path": "dia78_utt0.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5076 + }, + { + "path": "dia195_utt4.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5077 + }, + { + "path": "dia137_utt2.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5078 + }, + { + "path": "dia120_utt12.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5079 + }, + { + "path": "dia261_utt7.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5080 + }, + { + "path": "dia11_utt3.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5081 + }, + { + "path": "dia30_utt4.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5082 + }, + { + "path": "dia76_utt11.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5083 + }, + { + "path": "dia49_utt5.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5084 + }, + { + "path": "dia222_utt1.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5085 + }, + { + "path": "dia164_utt6.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5086 + }, + { + "path": "dia204_utt2.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5087 + }, + { + "path": "dia184_utt5.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5088 + }, + { + "path": "dia175_utt18.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5089 + }, + { + "path": "dia184_utt3.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5090 + }, + { + "path": "dia240_utt9.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5091 + }, + { + "path": "dia198_utt9.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5092 + }, + { + "path": "dia192_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5093 + }, + { + "path": "dia99_utt2.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5094 + }, + { + "path": "dia20_utt2.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5095 + }, + { + "path": "dia210_utt2.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5096 + }, + { + "path": "dia57_utt0.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5097 + }, + { + "path": "dia169_utt4.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5098 + }, + { + "path": "dia77_utt4.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5099 + }, + { + "path": "dia199_utt7.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5100 + }, + { + "path": "dia144_utt1.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5101 + }, + { + "path": "dia279_utt4.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5102 + }, + { + "path": "dia120_utt13.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5103 + }, + { + "path": "dia208_utt13.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5104 + }, + { + "path": "dia20_utt6.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5105 + }, + { + "path": "dia26_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5106 + }, + { + "path": "dia201_utt10.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5107 + }, + { + "path": "dia111_utt5.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5108 + }, + { + "path": "dia56_utt3.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5109 + }, + { + "path": "dia147_utt5.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5110 + }, + { + "path": "dia266_utt3.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5111 + }, + { + "path": "dia76_utt17.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5112 + }, + { + "path": "dia41_utt12.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5113 + }, + { + "path": "dia14_utt6.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5114 + }, + { + "path": "dia49_utt14.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5115 + }, + { + "path": "dia125_utt8.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5116 + }, + { + "path": "dia231_utt13.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5117 + }, + { + "path": "dia9_utt13.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5118 + }, + { + "path": "dia95_utt13.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5119 + }, + { + "path": "dia221_utt2.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5120 + }, + { + "path": "dia20_utt7.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5121 + }, + { + "path": "dia121_utt4.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5122 + }, + { + "path": "dia266_utt11.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5123 + }, + { + "path": "dia278_utt8.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5124 + }, + { + "path": "dia59_utt0.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5125 + }, + { + "path": "dia169_utt5.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5126 + }, + { + "path": "dia140_utt14.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5127 + }, + { + "path": "dia48_utt6.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5128 + }, + { + "path": "dia77_utt1.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5129 + }, + { + "path": "dia212_utt0.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5130 + }, + { + "path": "dia140_utt5.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5131 + }, + { + "path": "dia32_utt8.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5132 + }, + { + "path": "dia199_utt12.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5133 + }, + { + "path": "dia109_utt1.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5134 + }, + { + "path": "dia210_utt4.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5135 + }, + { + "path": "dia97_utt6.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5136 + }, + { + "path": "dia120_utt11.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5137 + }, + { + "path": "dia56_utt11.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5138 + }, + { + "path": "dia153_utt4.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5139 + }, + { + "path": "dia119_utt1.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5140 + }, + { + "path": "dia268_utt10.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5141 + }, + { + "path": "dia9_utt3.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5142 + }, + { + "path": "dia86_utt3.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5143 + }, + { + "path": "dia100_utt19.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5144 + }, + { + "path": "dia163_utt2.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5145 + }, + { + "path": "dia237_utt4.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5146 + }, + { + "path": "dia254_utt11.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5147 + }, + { + "path": "dia102_utt3.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5148 + }, + { + "path": "dia278_utt6.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5149 + }, + { + "path": "dia49_utt4.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5150 + }, + { + "path": "dia157_utt4.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5151 + }, + { + "path": "dia210_utt6.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5152 + }, + { + "path": "dia175_utt4.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5153 + }, + { + "path": "dia45_utt3.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5154 + }, + { + "path": "dia88_utt1.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5155 + }, + { + "path": "dia110_utt3.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5156 + }, + { + "path": "dia57_utt4.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5157 + }, + { + "path": "dia16_utt0.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5158 + }, + { + "path": "dia279_utt9.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5159 + }, + { + "path": "dia250_utt7.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5160 + }, + { + "path": "dia267_utt10.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5161 + }, + { + "path": "dia152_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5162 + }, + { + "path": "dia48_utt0.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5163 + }, + { + "path": "dia146_utt6.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5164 + }, + { + "path": "dia254_utt13.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5165 + }, + { + "path": "dia119_utt6.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5166 + }, + { + "path": "dia199_utt10.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5167 + }, + { + "path": "dia279_utt3.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5168 + }, + { + "path": "dia193_utt1.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5169 + }, + { + "path": "dia96_utt17.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5170 + }, + { + "path": "dia232_utt8.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5171 + }, + { + "path": "dia70_utt11.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5172 + }, + { + "path": "dia98_utt7.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5173 + }, + { + "path": "dia71_utt5.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5174 + }, + { + "path": "dia102_utt6.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5175 + }, + { + "path": "dia125_utt10.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5176 + }, + { + "path": "dia157_utt0.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5177 + }, + { + "path": "dia6_utt1.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5178 + }, + { + "path": "dia125_utt19.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5179 + }, + { + "path": "dia239_utt1.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5180 + }, + { + "path": "dia113_utt10.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5181 + }, + { + "path": "dia132_utt14.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5182 + }, + { + "path": "dia217_utt3.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5183 + }, + { + "path": "dia15_utt3.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5184 + }, + { + "path": "dia257_utt1.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5185 + }, + { + "path": "dia64_utt6.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5186 + }, + { + "path": "dia88_utt2.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5187 + }, + { + "path": "dia56_utt0.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5188 + }, + { + "path": "dia64_utt11.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5189 + }, + { + "path": "dia193_utt4.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5190 + }, + { + "path": "dia232_utt5.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5191 + }, + { + "path": "dia169_utt11.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5192 + }, + { + "path": "dia111_utt14.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5193 + }, + { + "path": "dia188_utt3.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5194 + }, + { + "path": "dia146_utt0.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5195 + }, + { + "path": "dia215_utt0.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5196 + }, + { + "path": "dia203_utt5.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5197 + }, + { + "path": "dia32_utt6.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5198 + }, + { + "path": "dia126_utt14.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5199 + }, + { + "path": "dia149_utt2.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5200 + }, + { + "path": "dia83_utt10.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5201 + }, + { + "path": "dia232_utt4.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5202 + }, + { + "path": "dia29_utt4.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5203 + }, + { + "path": "dia199_utt2.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5204 + }, + { + "path": "dia125_utt17.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5205 + }, + { + "path": "dia92_utt5.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5206 + }, + { + "path": "dia208_utt6.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5207 + }, + { + "path": "dia217_utt11.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5208 + }, + { + "path": "dia163_utt0.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5209 + }, + { + "path": "dia220_utt1.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5210 + }, + { + "path": "dia22_utt18.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5211 + }, + { + "path": "dia226_utt0.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5212 + }, + { + "path": "dia252_utt3.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5213 + }, + { + "path": "dia229_utt6.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5214 + }, + { + "path": "dia83_utt2.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5215 + }, + { + "path": "dia168_utt11.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5216 + }, + { + "path": "dia165_utt0.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5217 + }, + { + "path": "dia202_utt6.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5218 + }, + { + "path": "dia201_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5219 + }, + { + "path": "dia174_utt12.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5220 + }, + { + "path": "dia86_utt7.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5221 + }, + { + "path": "dia3_utt1.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5222 + }, + { + "path": "dia139_utt6.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5223 + }, + { + "path": "dia103_utt0.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5224 + }, + { + "path": "dia160_utt8.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5225 + }, + { + "path": "dia192_utt3.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5226 + }, + { + "path": "dia279_utt14.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5227 + }, + { + "path": "dia68_utt12.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5228 + }, + { + "path": "dia278_utt1.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5229 + }, + { + "path": "dia104_utt0.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5230 + }, + { + "path": "dia47_utt1.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5231 + }, + { + "path": "dia215_utt4.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5232 + }, + { + "path": "dia85_utt2.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5233 + }, + { + "path": "dia161_utt10.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5234 + }, + { + "path": "dia255_utt4.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5235 + }, + { + "path": "dia17_utt27.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5236 + }, + { + "path": "dia185_utt1.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5237 + }, + { + "path": "dia51_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5238 + }, + { + "path": "dia22_utt7.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5239 + }, + { + "path": "dia241_utt5.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5240 + }, + { + "path": "dia76_utt13.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5241 + }, + { + "path": "dia96_utt0.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5242 + }, + { + "path": "dia173_utt5.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5243 + }, + { + "path": "dia5_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5244 + }, + { + "path": "dia270_utt0.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5245 + }, + { + "path": "dia29_utt11.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5246 + }, + { + "path": "dia10_utt1.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5247 + }, + { + "path": "dia96_utt8.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5248 + }, + { + "path": "dia263_utt4.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5249 + }, + { + "path": "dia97_utt8.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5250 + }, + { + "path": "dia14_utt0.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5251 + }, + { + "path": "dia45_utt6.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5252 + }, + { + "path": "dia205_utt6.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5253 + }, + { + "path": "dia78_utt9.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5254 + }, + { + "path": "dia112_utt21.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5255 + }, + { + "path": "dia242_utt2.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5256 + }, + { + "path": "dia174_utt15.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5257 + }, + { + "path": "dia116_utt10.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5258 + }, + { + "path": "dia180_utt4.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5259 + }, + { + "path": "dia53_utt2.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5260 + }, + { + "path": "dia1_utt0.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5261 + }, + { + "path": "dia228_utt9.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5262 + }, + { + "path": "dia140_utt3.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5263 + }, + { + "path": "dia205_utt1.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5264 + }, + { + "path": "dia137_utt13.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5265 + }, + { + "path": "dia203_utt9.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5266 + }, + { + "path": "dia33_utt5.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5267 + }, + { + "path": "dia110_utt1.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5268 + }, + { + "path": "dia244_utt15.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5269 + }, + { + "path": "dia99_utt5.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5270 + }, + { + "path": "dia124_utt7.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5271 + }, + { + "path": "dia65_utt6.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5272 + }, + { + "path": "dia160_utt2.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5273 + }, + { + "path": "dia191_utt6.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5274 + }, + { + "path": "dia73_utt6.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5275 + }, + { + "path": "dia211_utt3.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5276 + }, + { + "path": "dia184_utt12.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5277 + }, + { + "path": "dia225_utt4.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5278 + }, + { + "path": "dia124_utt5.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5279 + }, + { + "path": "dia229_utt3.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5280 + }, + { + "path": "dia107_utt12.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5281 + }, + { + "path": "dia245_utt2.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5282 + }, + { + "path": "dia48_utt2.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5283 + }, + { + "path": "dia53_utt3.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5284 + }, + { + "path": "dia25_utt16.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5285 + }, + { + "path": "dia261_utt1.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5286 + }, + { + "path": "dia255_utt7.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5287 + }, + { + "path": "dia76_utt7.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5288 + }, + { + "path": "dia176_utt1.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5289 + }, + { + "path": "dia139_utt9.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5290 + }, + { + "path": "dia271_utt4.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5291 + }, + { + "path": "dia43_utt5.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5292 + }, + { + "path": "dia245_utt0.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5293 + }, + { + "path": "dia60_utt6.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5294 + }, + { + "path": "dia22_utt12.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5295 + }, + { + "path": "dia240_utt11.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5296 + }, + { + "path": "dia42_utt3.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5297 + }, + { + "path": "dia252_utt7.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5298 + }, + { + "path": "dia100_utt1.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5299 + }, + { + "path": "dia108_utt5.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5300 + }, + { + "path": "dia131_utt1.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5301 + }, + { + "path": "dia22_utt21.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5302 + }, + { + "path": "dia208_utt12.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5303 + }, + { + "path": "dia15_utt5.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5304 + }, + { + "path": "dia12_utt3.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5305 + }, + { + "path": "dia47_utt11.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5306 + }, + { + "path": "dia264_utt0.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5307 + }, + { + "path": "dia74_utt3.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5308 + }, + { + "path": "dia125_utt0.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5309 + }, + { + "path": "dia153_utt0.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5310 + }, + { + "path": "dia231_utt7.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5311 + }, + { + "path": "dia147_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5312 + }, + { + "path": "dia253_utt5.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5313 + }, + { + "path": "dia133_utt15.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5314 + }, + { + "path": "dia146_utt1.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5315 + }, + { + "path": "dia171_utt3.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5316 + }, + { + "path": "dia99_utt6.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5317 + }, + { + "path": "dia237_utt6.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5318 + }, + { + "path": "dia42_utt1.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5319 + }, + { + "path": "dia90_utt1.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5320 + }, + { + "path": "dia52_utt0.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5321 + }, + { + "path": "dia237_utt3.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5322 + }, + { + "path": "dia155_utt0.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5323 + }, + { + "path": "dia100_utt2.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5324 + }, + { + "path": "dia52_utt2.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5325 + }, + { + "path": "dia190_utt4.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5326 + }, + { + "path": "dia207_utt0.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5327 + }, + { + "path": "dia126_utt10.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5328 + }, + { + "path": "dia135_utt4.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5329 + }, + { + "path": "dia163_utt3.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5330 + }, + { + "path": "dia241_utt4.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5331 + }, + { + "path": "dia216_utt0.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5332 + }, + { + "path": "dia254_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5333 + }, + { + "path": "dia49_utt8.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5334 + }, + { + "path": "dia123_utt7.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5335 + }, + { + "path": "dia165_utt1.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5336 + }, + { + "path": "dia243_utt5.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5337 + }, + { + "path": "dia255_utt8.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5338 + }, + { + "path": "dia71_utt10.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5339 + }, + { + "path": "dia178_utt4.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5340 + }, + { + "path": "dia2_utt8.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5341 + }, + { + "path": "dia231_utt3.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5342 + }, + { + "path": "dia38_utt3.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5343 + }, + { + "path": "dia111_utt11.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5344 + }, + { + "path": "dia214_utt3.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5345 + }, + { + "path": "dia16_utt9.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5346 + }, + { + "path": "dia137_utt11.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5347 + }, + { + "path": "dia102_utt8.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5348 + }, + { + "path": "dia182_utt1.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5349 + }, + { + "path": "dia125_utt20.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5350 + }, + { + "path": "dia58_utt8.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5351 + }, + { + "path": "dia259_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5352 + }, + { + "path": "dia192_utt8.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5353 + }, + { + "path": "dia133_utt10.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5354 + }, + { + "path": "dia213_utt3.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5355 + }, + { + "path": "dia41_utt3.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5356 + }, + { + "path": "dia183_utt8.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5357 + }, + { + "path": "dia29_utt9.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5358 + }, + { + "path": "dia56_utt10.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5359 + }, + { + "path": "dia72_utt13.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5360 + }, + { + "path": "dia266_utt2.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5361 + }, + { + "path": "dia223_utt3.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5362 + }, + { + "path": "dia130_utt15.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5363 + }, + { + "path": "dia63_utt0.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5364 + }, + { + "path": "dia233_utt4.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5365 + }, + { + "path": "dia37_utt1.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5366 + }, + { + "path": "dia128_utt3.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5367 + }, + { + "path": "dia125_utt2.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5368 + }, + { + "path": "dia204_utt3.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5369 + }, + { + "path": "dia95_utt12.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5370 + }, + { + "path": "dia54_utt3.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5371 + }, + { + "path": "dia173_utt18.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5372 + }, + { + "path": "dia25_utt0.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5373 + }, + { + "path": "dia12_utt14.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5374 + }, + { + "path": "dia99_utt4.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5375 + }, + { + "path": "dia116_utt6.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5376 + }, + { + "path": "dia95_utt7.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5377 + }, + { + "path": "dia223_utt4.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5378 + }, + { + "path": "dia223_utt6.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5379 + }, + { + "path": "dia65_utt9.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5380 + }, + { + "path": "dia169_utt2.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5381 + }, + { + "path": "dia35_utt3.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5382 + }, + { + "path": "dia123_utt8.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5383 + }, + { + "path": "dia1_utt2.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5384 + }, + { + "path": "dia225_utt2.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5385 + }, + { + "path": "dia222_utt0.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5386 + }, + { + "path": "dia244_utt17.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5387 + }, + { + "path": "dia210_utt7.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5388 + }, + { + "path": "dia213_utt6.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5389 + }, + { + "path": "dia90_utt5.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5390 + }, + { + "path": "dia233_utt1.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5391 + }, + { + "path": "dia218_utt1.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5392 + }, + { + "path": "dia34_utt12.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5393 + }, + { + "path": "dia166_utt8.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5394 + }, + { + "path": "dia25_utt7.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5395 + }, + { + "path": "dia111_utt13.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5396 + }, + { + "path": "dia155_utt5.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5397 + }, + { + "path": "dia56_utt9.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5398 + }, + { + "path": "dia128_utt5.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5399 + }, + { + "path": "dia6_utt0.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5400 + }, + { + "path": "dia188_utt5.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5401 + }, + { + "path": "dia100_utt9.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5402 + }, + { + "path": "dia20_utt4.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5403 + }, + { + "path": "dia175_utt2.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5404 + }, + { + "path": "dia278_utt13.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5405 + }, + { + "path": "dia41_utt2.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5406 + }, + { + "path": "dia70_utt4.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5407 + }, + { + "path": "dia125_utt22.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5408 + }, + { + "path": "dia244_utt2.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5409 + }, + { + "path": "dia121_utt0.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5410 + }, + { + "path": "dia147_utt7.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5411 + }, + { + "path": "dia78_utt1.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5412 + }, + { + "path": "dia233_utt10.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5413 + }, + { + "path": "dia100_utt7.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5414 + }, + { + "path": "dia259_utt9.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5415 + }, + { + "path": "dia131_utt8.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5416 + }, + { + "path": "dia222_utt13.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5417 + }, + { + "path": "dia12_utt18.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5418 + }, + { + "path": "dia132_utt4.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5419 + }, + { + "path": "dia68_utt11.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5420 + }, + { + "path": "dia110_utt4.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5421 + }, + { + "path": "dia190_utt2.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5422 + }, + { + "path": "dia122_utt0.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5423 + }, + { + "path": "dia37_utt3.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5424 + }, + { + "path": "dia58_utt5.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5425 + }, + { + "path": "dia130_utt5.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5426 + }, + { + "path": "dia173_utt13.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5427 + }, + { + "path": "dia97_utt9.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5428 + }, + { + "path": "dia218_utt0.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5429 + }, + { + "path": "dia112_utt4.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5430 + }, + { + "path": "dia47_utt12.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5431 + }, + { + "path": "dia126_utt4.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5432 + }, + { + "path": "dia217_utt4.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5433 + }, + { + "path": "dia70_utt12.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5434 + }, + { + "path": "dia0_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5435 + }, + { + "path": "dia259_utt7.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5436 + }, + { + "path": "dia116_utt8.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5437 + }, + { + "path": "dia49_utt1.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5438 + }, + { + "path": "dia17_utt13.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5439 + }, + { + "path": "dia71_utt2.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5440 + }, + { + "path": "dia178_utt7.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5441 + }, + { + "path": "dia72_utt10.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5442 + }, + { + "path": "dia18_utt2.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5443 + }, + { + "path": "dia108_utt0.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5444 + }, + { + "path": "dia12_utt10.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5445 + }, + { + "path": "dia73_utt8.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5446 + }, + { + "path": "dia188_utt7.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5447 + }, + { + "path": "dia242_utt4.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5448 + }, + { + "path": "dia152_utt0.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5449 + }, + { + "path": "dia198_utt5.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5450 + }, + { + "path": "dia254_utt3.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5451 + }, + { + "path": "dia49_utt10.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5452 + }, + { + "path": "dia178_utt5.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5453 + }, + { + "path": "dia34_utt2.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5454 + }, + { + "path": "dia97_utt1.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5455 + }, + { + "path": "dia144_utt6.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5456 + }, + { + "path": "dia72_utt6.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5457 + }, + { + "path": "dia158_utt1.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5458 + }, + { + "path": "dia147_utt3.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5459 + }, + { + "path": "dia34_utt11.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5460 + }, + { + "path": "dia64_utt4.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5461 + }, + { + "path": "dia88_utt8.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5462 + }, + { + "path": "dia117_utt1.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5463 + }, + { + "path": "dia169_utt6.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5464 + }, + { + "path": "dia132_utt13.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5465 + }, + { + "path": "dia261_utt8.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5466 + }, + { + "path": "dia107_utt5.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5467 + }, + { + "path": "dia68_utt7.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5468 + }, + { + "path": "dia25_utt13.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5469 + }, + { + "path": "dia75_utt4.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5470 + }, + { + "path": "dia91_utt0.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5471 + }, + { + "path": "dia72_utt15.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5472 + }, + { + "path": "dia102_utt1.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5473 + }, + { + "path": "dia41_utt1.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5474 + }, + { + "path": "dia140_utt6.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5475 + }, + { + "path": "dia140_utt10.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5476 + }, + { + "path": "dia124_utt14.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5477 + }, + { + "path": "dia109_utt3.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5478 + }, + { + "path": "dia244_utt14.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5479 + }, + { + "path": "dia68_utt4.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5480 + }, + { + "path": "dia9_utt12.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5481 + }, + { + "path": "dia194_utt3.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5482 + }, + { + "path": "dia9_utt8.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5483 + }, + { + "path": "dia7_utt5.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5484 + }, + { + "path": "dia269_utt1.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5485 + }, + { + "path": "dia17_utt8.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5486 + }, + { + "path": "dia250_utt1.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5487 + }, + { + "path": "dia58_utt1.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5488 + }, + { + "path": "dia140_utt16.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5489 + }, + { + "path": "dia238_utt4.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5490 + }, + { + "path": "dia22_utt10.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5491 + }, + { + "path": "dia47_utt13.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5492 + }, + { + "path": "dia237_utt9.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5493 + }, + { + "path": "dia178_utt6.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5494 + }, + { + "path": "dia88_utt11.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5495 + }, + { + "path": "dia155_utt7.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5496 + }, + { + "path": "dia53_utt12.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5497 + }, + { + "path": "dia42_utt4.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5498 + }, + { + "path": "dia241_utt9.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5499 + }, + { + "path": "dia237_utt11.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5500 + }, + { + "path": "dia230_utt1.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5501 + }, + { + "path": "dia95_utt4.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5502 + }, + { + "path": "dia44_utt2.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5503 + }, + { + "path": "dia279_utt11.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5504 + }, + { + "path": "dia78_utt2.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5505 + }, + { + "path": "dia199_utt11.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5506 + }, + { + "path": "dia113_utt9.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5507 + }, + { + "path": "dia77_utt3.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5508 + }, + { + "path": "dia239_utt6.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5509 + }, + { + "path": "dia263_utt2.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5510 + }, + { + "path": "dia125_utt6.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5511 + }, + { + "path": "dia147_utt8.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5512 + }, + { + "path": "dia142_utt0.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5513 + }, + { + "path": "dia233_utt11.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5514 + }, + { + "path": "dia55_utt2.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5515 + }, + { + "path": "dia53_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5516 + }, + { + "path": "dia230_utt6.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5517 + }, + { + "path": "dia236_utt0.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5518 + }, + { + "path": "dia17_utt29.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5519 + }, + { + "path": "dia16_utt4.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5520 + }, + { + "path": "dia186_utt3.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5521 + }, + { + "path": "dia65_utt3.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5522 + }, + { + "path": "dia183_utt6.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5523 + }, + { + "path": "dia157_utt6.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5524 + }, + { + "path": "dia86_utt2.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5525 + }, + { + "path": "dia5_utt5.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5526 + }, + { + "path": "dia112_utt3.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5527 + }, + { + "path": "dia186_utt0.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5528 + }, + { + "path": "dia67_utt3.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5529 + }, + { + "path": "dia2_utt0.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5530 + }, + { + "path": "dia246_utt12.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5531 + }, + { + "path": "dia219_utt1.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5532 + }, + { + "path": "dia199_utt6.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5533 + }, + { + "path": "dia96_utt10.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5534 + }, + { + "path": "dia68_utt9.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5535 + }, + { + "path": "dia154_utt19.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5536 + }, + { + "path": "dia262_utt14.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5537 + }, + { + "path": "dia76_utt6.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5538 + }, + { + "path": "dia228_utt11.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5539 + }, + { + "path": "dia110_utt6.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5540 + }, + { + "path": "dia231_utt12.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5541 + }, + { + "path": "dia67_utt0.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5542 + }, + { + "path": "dia57_utt1.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5543 + }, + { + "path": "dia227_utt0.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5544 + }, + { + "path": "dia259_utt13.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5545 + }, + { + "path": "dia47_utt8.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5546 + }, + { + "path": "dia198_utt12.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5547 + }, + { + "path": "dia210_utt1.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5548 + }, + { + "path": "dia92_utt3.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5549 + }, + { + "path": "dia122_utt2.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5550 + }, + { + "path": "dia133_utt5.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5551 + }, + { + "path": "dia184_utt2.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5552 + }, + { + "path": "dia191_utt9.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5553 + }, + { + "path": "dia12_utt11.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5554 + }, + { + "path": "dia81_utt0.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5555 + }, + { + "path": "dia28_utt12.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5556 + }, + { + "path": "dia221_utt1.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5557 + }, + { + "path": "dia43_utt0.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5558 + }, + { + "path": "dia182_utt4.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5559 + }, + { + "path": "dia244_utt0.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5560 + }, + { + "path": "dia71_utt0.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5561 + }, + { + "path": "dia73_utt9.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5562 + }, + { + "path": "dia73_utt0.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5563 + }, + { + "path": "dia219_utt0.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5564 + }, + { + "path": "dia237_utt2.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5565 + }, + { + "path": "dia12_utt2.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5566 + }, + { + "path": "dia86_utt0.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5567 + }, + { + "path": "dia42_utt0.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5568 + }, + { + "path": "dia96_utt19.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5569 + }, + { + "path": "dia258_utt12.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5570 + }, + { + "path": "dia64_utt14.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5571 + }, + { + "path": "dia234_utt10.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5572 + }, + { + "path": "dia49_utt0.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5573 + }, + { + "path": "dia169_utt21.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5574 + }, + { + "path": "dia189_utt6.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5575 + }, + { + "path": "dia232_utt1.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5576 + }, + { + "path": "dia65_utt4.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5577 + }, + { + "path": "dia195_utt11.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5578 + }, + { + "path": "dia39_utt2.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5579 + }, + { + "path": "dia100_utt11.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5580 + }, + { + "path": "dia226_utt2.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5581 + }, + { + "path": "dia94_utt1.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5582 + }, + { + "path": "dia166_utt13.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5583 + }, + { + "path": "dia153_utt3.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5584 + }, + { + "path": "dia70_utt21.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5585 + }, + { + "path": "dia259_utt10.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5586 + }, + { + "path": "dia162_utt0.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5587 + }, + { + "path": "dia24_utt3.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5588 + }, + { + "path": "dia124_utt3.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5589 + }, + { + "path": "dia168_utt2.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5590 + }, + { + "path": "dia16_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5591 + }, + { + "path": "dia1_utt1.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5592 + }, + { + "path": "dia266_utt9.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5593 + }, + { + "path": "dia119_utt0.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5594 + }, + { + "path": "dia195_utt9.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5595 + }, + { + "path": "dia164_utt7.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5596 + }, + { + "path": "dia137_utt0.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5597 + }, + { + "path": "dia15_utt4.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5598 + }, + { + "path": "dia66_utt4.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5599 + }, + { + "path": "dia183_utt3.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5600 + }, + { + "path": "dia154_utt4.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5601 + }, + { + "path": "dia188_utt6.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5602 + }, + { + "path": "dia131_utt4.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5603 + }, + { + "path": "dia279_utt5.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5604 + }, + { + "path": "dia120_utt1.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5605 + }, + { + "path": "dia73_utt7.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5606 + }, + { + "path": "dia260_utt2.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5607 + }, + { + "path": "dia21_utt4.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5608 + }, + { + "path": "dia149_utt0.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5609 + }, + { + "path": "dia120_utt10.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5610 + }, + { + "path": "dia134_utt1.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5611 + }, + { + "path": "dia263_utt3.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5612 + }, + { + "path": "dia230_utt7.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5613 + }, + { + "path": "dia12_utt5.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5614 + }, + { + "path": "dia223_utt2.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5615 + }, + { + "path": "dia263_utt8.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5616 + }, + { + "path": "dia76_utt1.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5617 + }, + { + "path": "dia173_utt11.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5618 + }, + { + "path": "dia258_utt3.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5619 + }, + { + "path": "dia144_utt9.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5620 + }, + { + "path": "dia120_utt15.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5621 + }, + { + "path": "dia70_utt14.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5622 + }, + { + "path": "dia259_utt8.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5623 + }, + { + "path": "dia143_utt0.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5624 + }, + { + "path": "dia96_utt11.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5625 + }, + { + "path": "dia160_utt1.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5626 + }, + { + "path": "dia167_utt9.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5627 + }, + { + "path": "dia255_utt3.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5628 + }, + { + "path": "dia139_utt8.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5629 + }, + { + "path": "dia64_utt12.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5630 + }, + { + "path": "dia135_utt1.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5631 + }, + { + "path": "dia76_utt8.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5632 + }, + { + "path": "dia117_utt3.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5633 + }, + { + "path": "dia41_utt7.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5634 + }, + { + "path": "dia258_utt7.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5635 + }, + { + "path": "dia145_utt6.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5636 + }, + { + "path": "dia201_utt11.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5637 + }, + { + "path": "dia13_utt3.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5638 + }, + { + "path": "dia120_utt7.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5639 + }, + { + "path": "dia173_utt2.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5640 + }, + { + "path": "dia153_utt6.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5641 + }, + { + "path": "dia242_utt0.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5642 + }, + { + "path": "dia258_utt0.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5643 + }, + { + "path": "dia123_utt14.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5644 + }, + { + "path": "dia163_utt1.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5645 + }, + { + "path": "dia88_utt6.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5646 + }, + { + "path": "dia251_utt1.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5647 + }, + { + "path": "dia196_utt10.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5648 + }, + { + "path": "dia60_utt3.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5649 + }, + { + "path": "dia131_utt3.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5650 + }, + { + "path": "dia212_utt5.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5651 + }, + { + "path": "dia226_utt6.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5652 + }, + { + "path": "dia49_utt7.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5653 + }, + { + "path": "dia177_utt0.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5654 + }, + { + "path": "dia279_utt1.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5655 + }, + { + "path": "dia96_utt20.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5656 + }, + { + "path": "dia254_utt12.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5657 + }, + { + "path": "dia34_utt13.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5658 + }, + { + "path": "dia274_utt3.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5659 + }, + { + "path": "dia170_utt0.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5660 + }, + { + "path": "dia255_utt1.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5661 + }, + { + "path": "dia188_utt1.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5662 + }, + { + "path": "dia69_utt5.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5663 + }, + { + "path": "dia113_utt6.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5664 + }, + { + "path": "dia151_utt2.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5665 + }, + { + "path": "dia17_utt32.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5666 + }, + { + "path": "dia96_utt4.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5667 + }, + { + "path": "dia54_utt6.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5668 + }, + { + "path": "dia55_utt0.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5669 + }, + { + "path": "dia254_utt14.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5670 + }, + { + "path": "dia208_utt8.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5671 + }, + { + "path": "dia252_utt18.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5672 + }, + { + "path": "dia190_utt1.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5673 + }, + { + "path": "dia137_utt5.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5674 + }, + { + "path": "dia93_utt10.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5675 + }, + { + "path": "dia113_utt4.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5676 + }, + { + "path": "dia198_utt4.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5677 + }, + { + "path": "dia119_utt5.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5678 + }, + { + "path": "dia270_utt12.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5679 + }, + { + "path": "dia173_utt14.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5680 + }, + { + "path": "dia39_utt4.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5681 + }, + { + "path": "dia158_utt5.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5682 + }, + { + "path": "dia65_utt7.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5683 + }, + { + "path": "dia9_utt4.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5684 + }, + { + "path": "dia2_utt4.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5685 + }, + { + "path": "dia96_utt13.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5686 + }, + { + "path": "dia174_utt3.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5687 + }, + { + "path": "dia1_utt3.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5688 + }, + { + "path": "dia96_utt9.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5689 + }, + { + "path": "dia254_utt4.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5690 + }, + { + "path": "dia12_utt6.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5691 + }, + { + "path": "dia274_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5692 + }, + { + "path": "dia115_utt4.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5693 + }, + { + "path": "dia185_utt4.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5694 + }, + { + "path": "dia90_utt0.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5695 + }, + { + "path": "dia115_utt3.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5696 + }, + { + "path": "dia69_utt1.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5697 + }, + { + "path": "dia183_utt5.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5698 + }, + { + "path": "dia139_utt3.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5699 + }, + { + "path": "dia169_utt9.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5700 + }, + { + "path": "dia122_utt3.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5701 + }, + { + "path": "dia102_utt4.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5702 + }, + { + "path": "dia78_utt5.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5703 + }, + { + "path": "dia228_utt6.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5704 + }, + { + "path": "dia259_utt0.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5705 + }, + { + "path": "dia10_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5706 + }, + { + "path": "dia222_utt7.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5707 + }, + { + "path": "dia68_utt2.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5708 + }, + { + "path": "dia192_utt9.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5709 + }, + { + "path": "dia132_utt11.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5710 + }, + { + "path": "dia17_utt23.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5711 + }, + { + "path": "dia25_utt3.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5712 + }, + { + "path": "dia73_utt1.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5713 + }, + { + "path": "dia250_utt6.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5714 + }, + { + "path": "dia68_utt14.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5715 + }, + { + "path": "dia54_utt1.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5716 + }, + { + "path": "dia168_utt18.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5717 + }, + { + "path": "dia153_utt11.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5718 + }, + { + "path": "dia147_utt10.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5719 + }, + { + "path": "dia175_utt10.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5720 + }, + { + "path": "dia260_utt7.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5721 + }, + { + "path": "dia223_utt0.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5722 + }, + { + "path": "dia196_utt16.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5723 + }, + { + "path": "dia34_utt4.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5724 + }, + { + "path": "dia178_utt9.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5725 + }, + { + "path": "dia208_utt1.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5726 + }, + { + "path": "dia58_utt10.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5727 + }, + { + "path": "dia256_utt0.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5728 + }, + { + "path": "dia6_utt2.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5729 + }, + { + "path": "dia278_utt7.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5730 + }, + { + "path": "dia267_utt8.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5731 + }, + { + "path": "dia80_utt6.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5732 + }, + { + "path": "dia95_utt6.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5733 + }, + { + "path": "dia180_utt9.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5734 + }, + { + "path": "dia48_utt5.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5735 + }, + { + "path": "dia74_utt1.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5736 + }, + { + "path": "dia262_utt1.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5737 + }, + { + "path": "dia143_utt5.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5738 + }, + { + "path": "dia231_utt9.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5739 + }, + { + "path": "dia260_utt8.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5740 + }, + { + "path": "dia222_utt11.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5741 + }, + { + "path": "dia103_utt2.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5742 + }, + { + "path": "dia230_utt3.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5743 + }, + { + "path": "dia100_utt16.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "angry", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5744 + }, + { + "path": "dia157_utt7.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5745 + }, + { + "path": "dia96_utt2.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5746 + }, + { + "path": "dia222_utt3.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5747 + }, + { + "path": "dia244_utt16.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5748 + }, + { + "path": "dia53_utt11.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5749 + }, + { + "path": "dia262_utt16.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5750 + }, + { + "path": "dia100_utt4.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5751 + }, + { + "path": "dia22_utt11.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5752 + }, + { + "path": "dia100_utt17.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5753 + }, + { + "path": "dia160_utt3.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5754 + }, + { + "path": "dia103_utt5.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5755 + }, + { + "path": "dia25_utt15.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5756 + }, + { + "path": "dia118_utt9.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5757 + }, + { + "path": "dia187_utt8.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5758 + }, + { + "path": "dia274_utt4.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5759 + }, + { + "path": "dia161_utt1.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5760 + }, + { + "path": "dia49_utt3.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5761 + }, + { + "path": "dia17_utt9.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5762 + }, + { + "path": "dia173_utt3.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5763 + }, + { + "path": "dia185_utt7.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5764 + }, + { + "path": "dia167_utt1.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5765 + }, + { + "path": "dia29_utt2.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5766 + }, + { + "path": "dia83_utt9.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5767 + }, + { + "path": "dia198_utt0.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5768 + }, + { + "path": "dia205_utt7.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5769 + }, + { + "path": "dia58_utt13.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5770 + }, + { + "path": "dia17_utt3.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5771 + }, + { + "path": "dia134_utt3.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5772 + }, + { + "path": "dia32_utt4.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5773 + }, + { + "path": "dia258_utt5.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5774 + }, + { + "path": "dia93_utt9.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5775 + }, + { + "path": "dia170_utt4.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5776 + }, + { + "path": "dia182_utt6.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5777 + }, + { + "path": "dia112_utt13.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5778 + }, + { + "path": "dia278_utt11.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5779 + }, + { + "path": "dia54_utt7.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5780 + }, + { + "path": "dia130_utt1.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5781 + }, + { + "path": "dia258_utt11.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5782 + }, + { + "path": "dia120_utt16.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5783 + }, + { + "path": "dia68_utt8.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5784 + }, + { + "path": "dia188_utt2.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5785 + }, + { + "path": "dia220_utt0.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5786 + }, + { + "path": "dia145_utt11.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5787 + }, + { + "path": "dia4_utt1.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5788 + }, + { + "path": "dia7_utt0.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "disgusted", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5789 + }, + { + "path": "dia263_utt6.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5790 + }, + { + "path": "dia15_utt2.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5791 + }, + { + "path": "dia70_utt0.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5792 + }, + { + "path": "dia78_utt6.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5793 + }, + { + "path": "dia42_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5794 + }, + { + "path": "dia208_utt3.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5795 + }, + { + "path": "dia195_utt1.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5796 + }, + { + "path": "dia73_utt10.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5797 + }, + { + "path": "dia153_utt2.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5798 + }, + { + "path": "dia117_utt4.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5799 + }, + { + "path": "dia54_utt11.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5800 + }, + { + "path": "dia140_utt8.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5801 + }, + { + "path": "dia83_utt6.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5802 + }, + { + "path": "dia262_utt7.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5803 + }, + { + "path": "dia72_utt7.wav", + "question": "What affective tone can you discern from the speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5804 + }, + { + "path": "dia49_utt9.wav", + "question": "What vibe do you get from the speaker's emotional delivery?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5805 + }, + { + "path": "dia115_utt1.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5806 + }, + { + "path": "dia120_utt2.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5807 + }, + { + "path": "dia270_utt10.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5808 + }, + { + "path": "dia80_utt3.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5809 + }, + { + "path": "dia85_utt0.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "disgusted", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5810 + }, + { + "path": "dia166_utt5.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5811 + }, + { + "path": "dia246_utt0.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5812 + }, + { + "path": "dia25_utt1.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5813 + }, + { + "path": "dia252_utt11.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5814 + }, + { + "path": "dia7_utt3.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5815 + }, + { + "path": "dia249_utt1.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5816 + }, + { + "path": "dia79_utt0.wav", + "question": "What is the speaker's emotional pitch as you hear it?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5817 + }, + { + "path": "dia208_utt4.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5818 + }, + { + "path": "dia64_utt16.wav", + "question": "How do you interpret the speaker's emotional intonation?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5819 + }, + { + "path": "dia41_utt9.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5820 + }, + { + "path": "dia146_utt2.wav", + "question": "What feeling does the speaker's voice seem to convey?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5821 + }, + { + "path": "dia168_utt0.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5822 + }, + { + "path": "dia221_utt4.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5823 + }, + { + "path": "dia189_utt7.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5824 + }, + { + "path": "dia106_utt0.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5825 + }, + { + "path": "dia229_utt7.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5826 + }, + { + "path": "dia4_utt5.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5827 + }, + { + "path": "dia121_utt2.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5828 + }, + { + "path": "dia114_utt0.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5829 + }, + { + "path": "dia10_utt0.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5830 + }, + { + "path": "dia34_utt7.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5831 + }, + { + "path": "dia102_utt2.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5832 + }, + { + "path": "dia222_utt10.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5833 + }, + { + "path": "dia189_utt9.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5834 + }, + { + "path": "dia116_utt5.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5835 + }, + { + "path": "dia111_utt1.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5836 + }, + { + "path": "dia60_utt2.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5837 + }, + { + "path": "dia107_utt3.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5838 + }, + { + "path": "dia189_utt2.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5839 + }, + { + "path": "dia269_utt2.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5840 + }, + { + "path": "dia1_utt4.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5841 + }, + { + "path": "dia244_utt21.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5842 + }, + { + "path": "dia217_utt6.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5843 + }, + { + "path": "dia98_utt2.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5844 + }, + { + "path": "dia100_utt6.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5845 + }, + { + "path": "dia233_utt6.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5846 + }, + { + "path": "dia4_utt2.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5847 + }, + { + "path": "dia259_utt1.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5848 + }, + { + "path": "dia16_utt7.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5849 + }, + { + "path": "dia208_utt9.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5850 + }, + { + "path": "dia34_utt10.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "sad", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5851 + }, + { + "path": "dia258_utt4.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5852 + }, + { + "path": "dia250_utt3.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "neutral", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5853 + }, + { + "path": "dia45_utt7.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5854 + }, + { + "path": "dia196_utt2.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5855 + }, + { + "path": "dia202_utt5.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5856 + }, + { + "path": "dia76_utt4.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5857 + }, + { + "path": "dia110_utt5.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "surprised", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5858 + }, + { + "path": "dia28_utt16.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5859 + }, + { + "path": "dia87_utt1.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5860 + }, + { + "path": "dia196_utt9.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5861 + }, + { + "path": "dia90_utt4.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5862 + }, + { + "path": "dia167_utt3.wav", + "question": "What emotional atmosphere does the speech create?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5863 + }, + { + "path": "dia80_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5864 + }, + { + "path": "dia184_utt1.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5865 + }, + { + "path": "dia69_utt0.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5866 + }, + { + "path": "dia25_utt9.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5867 + }, + { + "path": "dia199_utt5.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5868 + }, + { + "path": "dia161_utt7.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "sad", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5869 + }, + { + "path": "dia28_utt6.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5870 + }, + { + "path": "dia152_utt4.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5871 + }, + { + "path": "dia278_utt5.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5872 + }, + { + "path": "dia233_utt9.wav", + "question": "What emotional message is delivered in the speech?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5873 + }, + { + "path": "dia90_utt3.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5874 + }, + { + "path": "dia17_utt22.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5875 + }, + { + "path": "dia22_utt15.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5876 + }, + { + "path": "dia174_utt5.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5877 + }, + { + "path": "dia131_utt6.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5878 + }, + { + "path": "dia78_utt8.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5879 + }, + { + "path": "dia22_utt0.wav", + "question": "How would you categorize the speaker's emotional state?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5880 + }, + { + "path": "dia120_utt8.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5881 + }, + { + "path": "dia153_utt9.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5882 + }, + { + "path": "dia100_utt3.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5883 + }, + { + "path": "dia89_utt2.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5884 + }, + { + "path": "dia170_utt1.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5885 + }, + { + "path": "dia45_utt4.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5886 + }, + { + "path": "dia138_utt3.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "angry", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5887 + }, + { + "path": "dia165_utt3.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "happy", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5888 + }, + { + "path": "dia160_utt11.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5889 + }, + { + "path": "dia273_utt1.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5890 + }, + { + "path": "dia71_utt12.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5891 + }, + { + "path": "dia147_utt0.wav", + "question": "Which emotion does the speaker's language evoke?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5892 + }, + { + "path": "dia113_utt12.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5893 + }, + { + "path": "dia146_utt3.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5894 + }, + { + "path": "dia9_utt10.wav", + "question": "What is the tone of emotion in the speaker's message?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5895 + }, + { + "path": "dia213_utt2.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "happy", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5896 + }, + { + "path": "dia154_utt3.wav", + "question": "In what emotional light do you perceive the speaker's delivery?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5897 + }, + { + "path": "dia156_utt1.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5898 + }, + { + "path": "dia126_utt8.wav", + "question": "How would you label the speaker's emotional inflection?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "angry", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5899 + }, + { + "path": "dia137_utt1.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5900 + }, + { + "path": "dia100_utt5.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5901 + }, + { + "path": "dia113_utt3.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5902 + }, + { + "path": "dia122_utt1.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "happy", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5903 + }, + { + "path": "dia204_utt6.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5904 + }, + { + "path": "dia241_utt2.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5905 + }, + { + "path": "dia76_utt16.wav", + "question": "What emotion is primarily conveyed in the speech?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5906 + }, + { + "path": "dia155_utt1.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5907 + }, + { + "path": "dia68_utt13.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5908 + }, + { + "path": "dia124_utt9.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "surprised", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5909 + }, + { + "path": "dia175_utt6.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5910 + }, + { + "path": "dia22_utt6.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5911 + }, + { + "path": "dia251_utt3.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "sad", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5912 + }, + { + "path": "dia21_utt6.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5913 + }, + { + "path": "dia17_utt15.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5914 + }, + { + "path": "dia125_utt9.wav", + "question": "What emotional undertone can you detect in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5915 + }, + { + "path": "dia213_utt5.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "fearful", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5916 + }, + { + "path": "dia153_utt10.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5917 + }, + { + "path": "dia98_utt9.wav", + "question": "What sentiment is the speaker predominantly expressing?", + "choice_a": "happy", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5918 + }, + { + "path": "dia182_utt5.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5919 + }, + { + "path": "dia144_utt10.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "neutral", + "choice_b": "sad", + "choice_c": "fearful", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5920 + }, + { + "path": "dia270_utt2.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5921 + }, + { + "path": "dia244_utt3.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5922 + }, + { + "path": "dia68_utt1.wav", + "question": "What emotion do you pick up from the speaker's verbal cues?", + "choice_a": "happy", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5923 + }, + { + "path": "dia229_utt0.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5924 + }, + { + "path": "dia196_utt5.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5925 + }, + { + "path": "dia61_utt1.wav", + "question": "What is the primary emotional impression you get from the speech?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5926 + }, + { + "path": "dia17_utt12.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5927 + }, + { + "path": "dia14_utt3.wav", + "question": "Which feeling does the speaker's tone best represent?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5928 + }, + { + "path": "dia108_utt7.wav", + "question": "What is the prevailing emotional tone of the speech?", + "choice_a": "fearful", + "choice_b": "sad", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5929 + }, + { + "path": "dia226_utt8.wav", + "question": "What underlying emotion do you detect in the speaker's expression?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5930 + }, + { + "path": "dia111_utt12.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "happy", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5931 + }, + { + "path": "dia257_utt3.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5932 + }, + { + "path": "dia230_utt8.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "surprised", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5933 + }, + { + "path": "dia70_utt16.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5934 + }, + { + "path": "dia144_utt0.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "fearful", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5935 + }, + { + "path": "dia195_utt7.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5936 + }, + { + "path": "dia99_utt9.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "sad", + "choice_b": "fearful", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5937 + }, + { + "path": "dia243_utt3.wav", + "question": "Which sentiment is the speaker communicating through their speech?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5938 + }, + { + "path": "dia196_utt4.wav", + "question": "What affective state does the speaker seem to be in?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "surprised", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5939 + }, + { + "path": "dia228_utt12.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "angry", + "choice_b": "surprised", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5940 + }, + { + "path": "dia33_utt8.wav", + "question": "How is the speaker's emotional stance reflected in their speech?", + "choice_a": "disgusted", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5941 + }, + { + "path": "dia161_utt11.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "surprised", + "choice_b": "fearful", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5942 + }, + { + "path": "dia56_utt5.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "surprised", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5943 + }, + { + "path": "dia33_utt1.wav", + "question": "Which affective label would you give to the speech's tone?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "surprised", + "choice_d": "sad", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5944 + }, + { + "path": "dia34_utt5.wav", + "question": "What is the emotional essence of the speaker's delivery?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "disgusted", + "choice_d": "surprised", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5945 + }, + { + "path": "dia95_utt19.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5946 + }, + { + "path": "dia123_utt2.wav", + "question": "Which of these emotions reflects the speech's tone?", + "choice_a": "happy", + "choice_b": "disgusted", + "choice_c": "neutral", + "choice_d": "angry", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5947 + }, + { + "path": "dia240_utt4.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5948 + }, + { + "path": "dia263_utt5.wav", + "question": "What is the speaker's mood as perceived through their speech?", + "choice_a": "happy", + "choice_b": "fearful", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5949 + }, + { + "path": "dia53_utt1.wav", + "question": "What is the dominant emotion expressed in the speech?", + "choice_a": "fearful", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5950 + }, + { + "path": "dia254_utt2.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "disgusted", + "choice_b": "happy", + "choice_c": "fearful", + "choice_d": "sad", + "answer_gt": "disgusted", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5951 + }, + { + "path": "dia95_utt2.wav", + "question": "What is the emotional color of the speaker's rhetoric?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5952 + }, + { + "path": "dia174_utt17.wav", + "question": "How does the speaker's emotion resonate with you?", + "choice_a": "angry", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5953 + }, + { + "path": "dia48_utt1.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "angry", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5954 + }, + { + "path": "dia193_utt3.wav", + "question": "What feeling pervades the speaker's discourse?", + "choice_a": "neutral", + "choice_b": "angry", + "choice_c": "surprised", + "choice_d": "fearful", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5955 + }, + { + "path": "dia12_utt4.wav", + "question": "What emotional tenor do you perceive in the speaker's voice?", + "choice_a": "disgusted", + "choice_b": "fearful", + "choice_c": "sad", + "choice_d": "angry", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5956 + }, + { + "path": "dia99_utt7.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "disgusted", + "choice_c": "surprised", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5957 + }, + { + "path": "dia239_utt2.wav", + "question": "What emotion is at the forefront of the speaker's words?", + "choice_a": "neutral", + "choice_b": "surprised", + "choice_c": "sad", + "choice_d": "fearful", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5958 + }, + { + "path": "dia271_utt0.wav", + "question": "What is your interpretation of the speaker's emotional timbre?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5959 + }, + { + "path": "dia22_utt1.wav", + "question": "What emotional vibe is conveyed by the speaker's words?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "neutral", + "choice_d": "disgusted", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5960 + }, + { + "path": "dia96_utt18.wav", + "question": "What emotional quality do you associate with the speech?", + "choice_a": "sad", + "choice_b": "surprised", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5961 + }, + { + "path": "dia228_utt5.wav", + "question": "Which emotion resonates with the content of the speech?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "happy", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5962 + }, + { + "path": "dia223_utt1.wav", + "question": "Which emotional nuance is evident in the speaker's speech?", + "choice_a": "angry", + "choice_b": "happy", + "choice_c": "disgusted", + "choice_d": "fearful", + "answer_gt": "fearful", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5963 + }, + { + "path": "dia34_utt17.wav", + "question": "Which of these emotions matches the speaker's sentiments?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "fearful", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5964 + }, + { + "path": "dia12_utt7.wav", + "question": "How would you characterize the emotional quality of the speech?", + "choice_a": "neutral", + "choice_b": "disgusted", + "choice_c": "angry", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5965 + }, + { + "path": "dia230_utt4.wav", + "question": "How would you sum up the speaker's emotional state?", + "choice_a": "surprised", + "choice_b": "angry", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5966 + }, + { + "path": "dia65_utt1.wav", + "question": "How does the speaker's emotional expression come across?", + "choice_a": "surprised", + "choice_b": "neutral", + "choice_c": "sad", + "choice_d": "happy", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5967 + }, + { + "path": "dia275_utt1.wav", + "question": "What mood is most prominent in the speaker's delivery?", + "choice_a": "sad", + "choice_b": "angry", + "choice_c": "fearful", + "choice_d": "neutral", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5968 + }, + { + "path": "dia83_utt5.wav", + "question": "What sentiment comes through most clearly in the speech?", + "choice_a": "angry", + "choice_b": "neutral", + "choice_c": "surprised", + "choice_d": "disgusted", + "answer_gt": "angry", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5969 + }, + { + "path": "dia118_utt2.wav", + "question": "Which emotion do you think the speech embodies?", + "choice_a": "fearful", + "choice_b": "surprised", + "choice_c": "disgusted", + "choice_d": "angry", + "answer_gt": "surprised", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5970 + }, + { + "path": "dia1_utt7.wav", + "question": "How do you feel about the speaker's emotional tone?", + "choice_a": "disgusted", + "choice_b": "sad", + "choice_c": "neutral", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5971 + }, + { + "path": "dia29_utt8.wav", + "question": "Which emotional description fits the speaker's message best?", + "choice_a": "happy", + "choice_b": "angry", + "choice_c": "neutral", + "choice_d": "sad", + "answer_gt": "neutral", + "task_name": "Speaker_Emotion_Recontion", + "dataset_name": "meld", + "uniq_id": 5972 + }, + { + "path": "common_voice_en_22159151.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5973 + }, + { + "path": "common_voice_en_17297655.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5974 + }, + { + "path": "common_voice_en_17152441.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5975 + }, + { + "path": "common_voice_en_36862699.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5976 + }, + { + "path": "common_voice_en_19214185.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5977 + }, + { + "path": "common_voice_en_700697.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5978 + }, + { + "path": "common_voice_en_17281452.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5979 + }, + { + "path": "common_voice_en_36544758.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5980 + }, + { + "path": "common_voice_en_31041412.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5981 + }, + { + "path": "common_voice_en_459460.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5982 + }, + { + "path": "common_voice_en_136405.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5983 + }, + { + "path": "common_voice_en_34693.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5984 + }, + { + "path": "common_voice_en_31070494.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5985 + }, + { + "path": "common_voice_en_18012943.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5986 + }, + { + "path": "common_voice_en_21099981.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5987 + }, + { + "path": "common_voice_en_23828198.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5988 + }, + { + "path": "common_voice_en_607195.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5989 + }, + { + "path": "common_voice_en_522868.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5990 + }, + { + "path": "common_voice_en_20182238.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5991 + }, + { + "path": "common_voice_en_19935427.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5992 + }, + { + "path": "common_voice_en_18857334.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5993 + }, + { + "path": "common_voice_en_31562356.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5994 + }, + { + "path": "common_voice_en_20678930.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5995 + }, + { + "path": "common_voice_en_19808302.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5996 + }, + { + "path": "common_voice_en_19954794.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5997 + }, + { + "path": "common_voice_en_21761955.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5998 + }, + { + "path": "common_voice_en_19050519.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 5999 + }, + { + "path": "common_voice_en_20911760.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6000 + }, + { + "path": "common_voice_en_22605409.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6001 + }, + { + "path": "common_voice_en_18255019.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6002 + }, + { + "path": "common_voice_en_26209043.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6003 + }, + { + "path": "common_voice_en_36068063.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6004 + }, + { + "path": "common_voice_en_21883920.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6005 + }, + { + "path": "common_voice_en_17289324.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6006 + }, + { + "path": "common_voice_en_17259866.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6007 + }, + { + "path": "common_voice_en_187602.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6008 + }, + { + "path": "common_voice_en_22450351.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6009 + }, + { + "path": "common_voice_en_16706012.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6010 + }, + { + "path": "common_voice_en_31682687.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6011 + }, + { + "path": "common_voice_en_696712.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6012 + }, + { + "path": "common_voice_en_26966176.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6013 + }, + { + "path": "common_voice_en_17831791.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6014 + }, + { + "path": "common_voice_en_23854634.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6015 + }, + { + "path": "common_voice_en_17371751.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6016 + }, + { + "path": "common_voice_en_20781208.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6017 + }, + { + "path": "common_voice_en_21350769.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6018 + }, + { + "path": "common_voice_en_19938759.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6019 + }, + { + "path": "common_voice_en_20258571.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6020 + }, + { + "path": "common_voice_en_19693120.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6021 + }, + { + "path": "common_voice_en_35832645.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6022 + }, + { + "path": "common_voice_en_31650547.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6023 + }, + { + "path": "common_voice_en_31893855.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6024 + }, + { + "path": "common_voice_en_20172863.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6025 + }, + { + "path": "common_voice_en_21387748.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6026 + }, + { + "path": "common_voice_en_31266585.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6027 + }, + { + "path": "common_voice_en_19982627.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6028 + }, + { + "path": "common_voice_en_442796.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6029 + }, + { + "path": "common_voice_en_57396.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6030 + }, + { + "path": "common_voice_en_19212042.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6031 + }, + { + "path": "common_voice_en_36290114.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6032 + }, + { + "path": "common_voice_en_19736952.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6033 + }, + { + "path": "common_voice_en_17252413.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6034 + }, + { + "path": "common_voice_en_28807929.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6035 + }, + { + "path": "common_voice_en_24451253.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6036 + }, + { + "path": "common_voice_en_29540099.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6037 + }, + { + "path": "common_voice_en_35175851.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6038 + }, + { + "path": "common_voice_en_19707483.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6039 + }, + { + "path": "common_voice_en_20283484.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6040 + }, + { + "path": "common_voice_en_19078894.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6041 + }, + { + "path": "common_voice_en_17857149.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6042 + }, + { + "path": "common_voice_en_18332492.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6043 + }, + { + "path": "common_voice_en_23568648.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6044 + }, + { + "path": "common_voice_en_586565.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6045 + }, + { + "path": "common_voice_en_18541004.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6046 + }, + { + "path": "common_voice_en_21069613.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6047 + }, + { + "path": "common_voice_en_17747097.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6048 + }, + { + "path": "common_voice_en_532087.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6049 + }, + { + "path": "common_voice_en_19209768.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6050 + }, + { + "path": "common_voice_en_17758246.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6051 + }, + { + "path": "common_voice_en_322482.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6052 + }, + { + "path": "common_voice_en_26287634.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6053 + }, + { + "path": "common_voice_en_30585685.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6054 + }, + { + "path": "common_voice_en_23704834.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6055 + }, + { + "path": "common_voice_en_17151232.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6056 + }, + { + "path": "common_voice_en_21522616.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6057 + }, + { + "path": "common_voice_en_19752016.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6058 + }, + { + "path": "common_voice_en_15734801.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6059 + }, + { + "path": "common_voice_en_24472733.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6060 + }, + { + "path": "common_voice_en_31755671.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6061 + }, + { + "path": "common_voice_en_21264423.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6062 + }, + { + "path": "common_voice_en_16706008.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6063 + }, + { + "path": "common_voice_en_32642818.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6064 + }, + { + "path": "common_voice_en_37203803.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6065 + }, + { + "path": "common_voice_en_30717419.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6066 + }, + { + "path": "common_voice_en_26717744.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6067 + }, + { + "path": "common_voice_en_19553272.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6068 + }, + { + "path": "common_voice_en_2731682.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6069 + }, + { + "path": "common_voice_en_21403094.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6070 + }, + { + "path": "common_voice_en_30699821.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6071 + }, + { + "path": "common_voice_en_207024.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6072 + }, + { + "path": "common_voice_en_20644054.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6073 + }, + { + "path": "common_voice_en_17285876.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6074 + }, + { + "path": "common_voice_en_31728114.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6075 + }, + { + "path": "common_voice_en_367927.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6076 + }, + { + "path": "common_voice_en_31901950.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6077 + }, + { + "path": "common_voice_en_17730429.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6078 + }, + { + "path": "common_voice_en_17647993.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6079 + }, + { + "path": "common_voice_en_1742052.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6080 + }, + { + "path": "common_voice_en_29440889.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6081 + }, + { + "path": "common_voice_en_20405322.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6082 + }, + { + "path": "common_voice_en_19531689.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6083 + }, + { + "path": "common_voice_en_19412383.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6084 + }, + { + "path": "common_voice_en_19846920.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6085 + }, + { + "path": "common_voice_en_19826449.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6086 + }, + { + "path": "common_voice_en_20067906.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6087 + }, + { + "path": "common_voice_en_32538389.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6088 + }, + { + "path": "common_voice_en_32144882.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6089 + }, + { + "path": "common_voice_en_187059.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6090 + }, + { + "path": "common_voice_en_27386232.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6091 + }, + { + "path": "common_voice_en_32144886.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6092 + }, + { + "path": "common_voice_en_31661608.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6093 + }, + { + "path": "common_voice_en_18541003.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6094 + }, + { + "path": "common_voice_en_32117457.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6095 + }, + { + "path": "common_voice_en_23827068.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6096 + }, + { + "path": "common_voice_en_17295580.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6097 + }, + { + "path": "common_voice_en_17285890.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6098 + }, + { + "path": "common_voice_en_373478.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6099 + }, + { + "path": "common_voice_en_31637905.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6100 + }, + { + "path": "common_voice_en_18101571.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6101 + }, + { + "path": "common_voice_en_20290.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6102 + }, + { + "path": "common_voice_en_32101019.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6103 + }, + { + "path": "common_voice_en_17247324.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6104 + }, + { + "path": "common_voice_en_20239311.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6105 + }, + { + "path": "common_voice_en_18508474.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6106 + }, + { + "path": "common_voice_en_21300206.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6107 + }, + { + "path": "common_voice_en_18127728.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6108 + }, + { + "path": "common_voice_en_26287633.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6109 + }, + { + "path": "common_voice_en_17843117.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6110 + }, + { + "path": "common_voice_en_19672834.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6111 + }, + { + "path": "common_voice_en_18813644.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6112 + }, + { + "path": "common_voice_en_18303527.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6113 + }, + { + "path": "common_voice_en_24009284.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6114 + }, + { + "path": "common_voice_en_21567727.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6115 + }, + { + "path": "common_voice_en_29402882.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6116 + }, + { + "path": "common_voice_en_36526623.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6117 + }, + { + "path": "common_voice_en_18268920.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6118 + }, + { + "path": "common_voice_en_19951112.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6119 + }, + { + "path": "common_voice_en_687473.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6120 + }, + { + "path": "common_voice_en_20301609.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6121 + }, + { + "path": "common_voice_en_19553273.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6122 + }, + { + "path": "common_voice_en_33378405.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6123 + }, + { + "path": "common_voice_en_99416.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6124 + }, + { + "path": "common_voice_en_33290227.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6125 + }, + { + "path": "common_voice_en_24142666.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6126 + }, + { + "path": "common_voice_en_26129448.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6127 + }, + { + "path": "common_voice_en_19715449.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6128 + }, + { + "path": "common_voice_en_18593745.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6129 + }, + { + "path": "common_voice_en_690243.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6130 + }, + { + "path": "common_voice_en_23568649.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6131 + }, + { + "path": "common_voice_en_31692921.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6132 + }, + { + "path": "common_voice_en_19524811.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6133 + }, + { + "path": "common_voice_en_20508033.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6134 + }, + { + "path": "common_voice_en_22717228.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6135 + }, + { + "path": "common_voice_en_20178859.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6136 + }, + { + "path": "common_voice_en_31665253.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6137 + }, + { + "path": "common_voice_en_31573009.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6138 + }, + { + "path": "common_voice_en_19351145.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6139 + }, + { + "path": "common_voice_en_20644765.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6140 + }, + { + "path": "common_voice_en_30741424.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6141 + }, + { + "path": "common_voice_en_93238.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6142 + }, + { + "path": "common_voice_en_32280983.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6143 + }, + { + "path": "common_voice_en_24278995.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6144 + }, + { + "path": "common_voice_en_1564420.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6145 + }, + { + "path": "common_voice_en_17275001.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6146 + }, + { + "path": "common_voice_en_33085853.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6147 + }, + { + "path": "common_voice_en_1282108.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6148 + }, + { + "path": "common_voice_en_37228635.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6149 + }, + { + "path": "common_voice_en_34299867.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6150 + }, + { + "path": "common_voice_en_35798064.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6151 + }, + { + "path": "common_voice_en_6640943.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6152 + }, + { + "path": "common_voice_en_23749403.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6153 + }, + { + "path": "common_voice_en_34891889.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6154 + }, + { + "path": "common_voice_en_37103360.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6155 + }, + { + "path": "common_voice_en_23657133.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6156 + }, + { + "path": "common_voice_en_17417833.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6157 + }, + { + "path": "common_voice_en_19625460.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6158 + }, + { + "path": "common_voice_en_31676945.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6159 + }, + { + "path": "common_voice_en_35328182.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6160 + }, + { + "path": "common_voice_en_20009193.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6161 + }, + { + "path": "common_voice_en_18707936.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6162 + }, + { + "path": "common_voice_en_648458.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6163 + }, + { + "path": "common_voice_en_18283644.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6164 + }, + { + "path": "common_voice_en_32171092.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6165 + }, + { + "path": "common_voice_en_34321811.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6166 + }, + { + "path": "common_voice_en_24278994.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6167 + }, + { + "path": "common_voice_en_29346877.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6168 + }, + { + "path": "common_voice_en_23539705.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6169 + }, + { + "path": "common_voice_en_23907754.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6170 + }, + { + "path": "common_voice_en_19738882.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6171 + }, + { + "path": "common_voice_en_23814330.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6172 + }, + { + "path": "common_voice_en_19682043.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6173 + }, + { + "path": "common_voice_en_20489055.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6174 + }, + { + "path": "common_voice_en_19934255.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6175 + }, + { + "path": "common_voice_en_543775.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6176 + }, + { + "path": "common_voice_en_458608.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6177 + }, + { + "path": "common_voice_en_10410523.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6178 + }, + { + "path": "common_voice_en_30560272.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6179 + }, + { + "path": "common_voice_en_22802253.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6180 + }, + { + "path": "common_voice_en_18340487.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6181 + }, + { + "path": "common_voice_en_23914983.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6182 + }, + { + "path": "common_voice_en_19613456.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6183 + }, + { + "path": "common_voice_en_22754639.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6184 + }, + { + "path": "common_voice_en_31572632.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6185 + }, + { + "path": "common_voice_en_29440894.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6186 + }, + { + "path": "common_voice_en_19956821.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6187 + }, + { + "path": "common_voice_en_18969138.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6188 + }, + { + "path": "common_voice_en_19941596.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6189 + }, + { + "path": "common_voice_en_19756110.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6190 + }, + { + "path": "common_voice_en_20178860.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6191 + }, + { + "path": "common_voice_en_18725229.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6192 + }, + { + "path": "common_voice_en_18342344.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6193 + }, + { + "path": "common_voice_en_18828400.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6194 + }, + { + "path": "common_voice_en_27734888.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6195 + }, + { + "path": "common_voice_en_34989981.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6196 + }, + { + "path": "common_voice_en_32311207.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6197 + }, + { + "path": "common_voice_en_36576495.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6198 + }, + { + "path": "common_voice_en_24230074.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6199 + }, + { + "path": "common_voice_en_552059.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6200 + }, + { + "path": "common_voice_en_16665997.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6201 + }, + { + "path": "common_voice_en_19041344.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6202 + }, + { + "path": "common_voice_en_20047839.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6203 + }, + { + "path": "common_voice_en_627541.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6204 + }, + { + "path": "common_voice_en_25734384.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6205 + }, + { + "path": "common_voice_en_32573474.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6206 + }, + { + "path": "common_voice_en_19696555.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6207 + }, + { + "path": "common_voice_en_483208.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6208 + }, + { + "path": "common_voice_en_31534149.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6209 + }, + { + "path": "common_voice_en_31512831.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6210 + }, + { + "path": "common_voice_en_3792740.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6211 + }, + { + "path": "common_voice_en_276368.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6212 + }, + { + "path": "common_voice_en_19650893.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6213 + }, + { + "path": "common_voice_en_33344944.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6214 + }, + { + "path": "common_voice_en_22570760.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6215 + }, + { + "path": "common_voice_en_367926.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6216 + }, + { + "path": "common_voice_en_34298351.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6217 + }, + { + "path": "common_voice_en_512011.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6218 + }, + { + "path": "common_voice_en_31661492.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6219 + }, + { + "path": "common_voice_en_17779717.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6220 + }, + { + "path": "common_voice_en_20440161.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6221 + }, + { + "path": "common_voice_en_675201.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6222 + }, + { + "path": "common_voice_en_37269569.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6223 + }, + { + "path": "common_voice_en_19732218.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6224 + }, + { + "path": "common_voice_en_18124422.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6225 + }, + { + "path": "common_voice_en_20036344.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6226 + }, + { + "path": "common_voice_en_19611933.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6227 + }, + { + "path": "common_voice_en_32929058.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6228 + }, + { + "path": "common_voice_en_18770692.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6229 + }, + { + "path": "common_voice_en_17736913.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6230 + }, + { + "path": "common_voice_en_17936458.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6231 + }, + { + "path": "common_voice_en_27416441.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6232 + }, + { + "path": "common_voice_en_20236226.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6233 + }, + { + "path": "common_voice_en_4561565.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6234 + }, + { + "path": "common_voice_en_19964357.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6235 + }, + { + "path": "common_voice_en_20222177.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6236 + }, + { + "path": "common_voice_en_31657631.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6237 + }, + { + "path": "common_voice_en_31637920.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6238 + }, + { + "path": "common_voice_en_17250258.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6239 + }, + { + "path": "common_voice_en_31007974.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6240 + }, + { + "path": "common_voice_en_37118167.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6241 + }, + { + "path": "common_voice_en_17730790.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6242 + }, + { + "path": "common_voice_en_17902010.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6243 + }, + { + "path": "common_voice_en_16030401.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6244 + }, + { + "path": "common_voice_en_31536928.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6245 + }, + { + "path": "common_voice_en_211447.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6246 + }, + { + "path": "common_voice_en_20900715.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6247 + }, + { + "path": "common_voice_en_25126314.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6248 + }, + { + "path": "common_voice_en_22910972.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6249 + }, + { + "path": "common_voice_en_19680459.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6250 + }, + { + "path": "common_voice_en_18354220.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6251 + }, + { + "path": "common_voice_en_32875324.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6252 + }, + { + "path": "common_voice_en_26287636.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6253 + }, + { + "path": "common_voice_en_17624849.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6254 + }, + { + "path": "common_voice_en_31866896.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6255 + }, + { + "path": "common_voice_en_31811127.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6256 + }, + { + "path": "common_voice_en_16665998.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6257 + }, + { + "path": "common_voice_en_20275070.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6258 + }, + { + "path": "common_voice_en_36455480.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6259 + }, + { + "path": "common_voice_en_18166798.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6260 + }, + { + "path": "common_voice_en_17249432.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6261 + }, + { + "path": "common_voice_en_13483785.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6262 + }, + { + "path": "common_voice_en_18306538.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6263 + }, + { + "path": "common_voice_en_21702346.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6264 + }, + { + "path": "common_voice_en_17281451.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6265 + }, + { + "path": "common_voice_en_28458841.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6266 + }, + { + "path": "common_voice_en_508299.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6267 + }, + { + "path": "common_voice_en_17904706.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6268 + }, + { + "path": "common_voice_en_31586709.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6269 + }, + { + "path": "common_voice_en_26746245.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6270 + }, + { + "path": "common_voice_en_21321827.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6271 + }, + { + "path": "common_voice_en_19011343.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6272 + }, + { + "path": "common_voice_en_70723.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6273 + }, + { + "path": "common_voice_en_22951225.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6274 + }, + { + "path": "common_voice_en_19789702.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6275 + }, + { + "path": "common_voice_en_356183.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6276 + }, + { + "path": "common_voice_en_19748963.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6277 + }, + { + "path": "common_voice_en_15735758.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6278 + }, + { + "path": "common_voice_en_21754598.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6279 + }, + { + "path": "common_voice_en_33193621.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6280 + }, + { + "path": "common_voice_en_18508473.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6281 + }, + { + "path": "common_voice_en_19011864.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6282 + }, + { + "path": "common_voice_en_18547306.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6283 + }, + { + "path": "common_voice_en_24261278.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6284 + }, + { + "path": "common_voice_en_18712690.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6285 + }, + { + "path": "common_voice_en_1531516.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6286 + }, + { + "path": "common_voice_en_18758347.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6287 + }, + { + "path": "common_voice_en_17289296.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6288 + }, + { + "path": "common_voice_en_20009021.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6289 + }, + { + "path": "common_voice_en_18977842.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6290 + }, + { + "path": "common_voice_en_33452651.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6291 + }, + { + "path": "common_voice_en_20688778.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6292 + }, + { + "path": "common_voice_en_16659359.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6293 + }, + { + "path": "common_voice_en_21568217.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6294 + }, + { + "path": "common_voice_en_21340519.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6295 + }, + { + "path": "common_voice_en_18485414.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6296 + }, + { + "path": "common_voice_en_18882317.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6297 + }, + { + "path": "common_voice_en_37037383.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6298 + }, + { + "path": "common_voice_en_177712.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6299 + }, + { + "path": "common_voice_en_17728545.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6300 + }, + { + "path": "common_voice_en_28458839.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6301 + }, + { + "path": "common_voice_en_19496792.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6302 + }, + { + "path": "common_voice_en_17727328.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6303 + }, + { + "path": "common_voice_en_31821767.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6304 + }, + { + "path": "common_voice_en_20688737.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6305 + }, + { + "path": "common_voice_en_18254939.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6306 + }, + { + "path": "common_voice_en_25257209.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6307 + }, + { + "path": "common_voice_en_18238070.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6308 + }, + { + "path": "common_voice_en_6225638.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6309 + }, + { + "path": "common_voice_en_18183180.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6310 + }, + { + "path": "common_voice_en_20631954.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6311 + }, + { + "path": "common_voice_en_19477710.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6312 + }, + { + "path": "common_voice_en_18854899.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6313 + }, + { + "path": "common_voice_en_27734771.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6314 + }, + { + "path": "common_voice_en_33443861.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6315 + }, + { + "path": "common_voice_en_19214190.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6316 + }, + { + "path": "common_voice_en_18758926.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6317 + }, + { + "path": "common_voice_en_19100981.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6318 + }, + { + "path": "common_voice_en_30981010.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6319 + }, + { + "path": "common_voice_en_31578826.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6320 + }, + { + "path": "common_voice_en_18304023.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6321 + }, + { + "path": "common_voice_en_23489871.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6322 + }, + { + "path": "common_voice_en_17269127.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6323 + }, + { + "path": "common_voice_en_17354534.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6324 + }, + { + "path": "common_voice_en_32501175.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6325 + }, + { + "path": "common_voice_en_20221794.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6326 + }, + { + "path": "common_voice_en_6640942.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6327 + }, + { + "path": "common_voice_en_32100835.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6328 + }, + { + "path": "common_voice_en_21858027.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6329 + }, + { + "path": "common_voice_en_20662756.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6330 + }, + { + "path": "common_voice_en_18730818.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6331 + }, + { + "path": "common_voice_en_32929052.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6332 + }, + { + "path": "common_voice_en_534195.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6333 + }, + { + "path": "common_voice_en_24427130.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6334 + }, + { + "path": "common_voice_en_36499457.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6335 + }, + { + "path": "common_voice_en_1323349.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6336 + }, + { + "path": "common_voice_en_28875056.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6337 + }, + { + "path": "common_voice_en_19685105.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6338 + }, + { + "path": "common_voice_en_23801930.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6339 + }, + { + "path": "common_voice_en_17928241.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6340 + }, + { + "path": "common_voice_en_18607767.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6341 + }, + { + "path": "common_voice_en_31618144.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6342 + }, + { + "path": "common_voice_en_18762118.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6343 + }, + { + "path": "common_voice_en_205395.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6344 + }, + { + "path": "common_voice_en_20688728.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6345 + }, + { + "path": "common_voice_en_17283107.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6346 + }, + { + "path": "common_voice_en_19015151.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6347 + }, + { + "path": "common_voice_en_21786428.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6348 + }, + { + "path": "common_voice_en_19288096.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6349 + }, + { + "path": "common_voice_en_573140.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6350 + }, + { + "path": "common_voice_en_32875328.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6351 + }, + { + "path": "common_voice_en_23686301.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6352 + }, + { + "path": "common_voice_en_18017628.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6353 + }, + { + "path": "common_voice_en_32744340.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6354 + }, + { + "path": "common_voice_en_17895581.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6355 + }, + { + "path": "common_voice_en_18848277.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6356 + }, + { + "path": "common_voice_en_22559121.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6357 + }, + { + "path": "common_voice_en_30713100.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6358 + }, + { + "path": "common_voice_en_19169722.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6359 + }, + { + "path": "common_voice_en_32366593.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6360 + }, + { + "path": "common_voice_en_22975991.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6361 + }, + { + "path": "common_voice_en_25884084.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6362 + }, + { + "path": "common_voice_en_21685243.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6363 + }, + { + "path": "common_voice_en_19945491.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6364 + }, + { + "path": "common_voice_en_20854413.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6365 + }, + { + "path": "common_voice_en_19617110.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6366 + }, + { + "path": "common_voice_en_37083044.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6367 + }, + { + "path": "common_voice_en_25031275.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6368 + }, + { + "path": "common_voice_en_19962548.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6369 + }, + { + "path": "common_voice_en_28807933.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6370 + }, + { + "path": "common_voice_en_17250887.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6371 + }, + { + "path": "common_voice_en_35582638.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6372 + }, + { + "path": "common_voice_en_23682212.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6373 + }, + { + "path": "common_voice_en_20688736.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6374 + }, + { + "path": "common_voice_en_25337475.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6375 + }, + { + "path": "common_voice_en_595873.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6376 + }, + { + "path": "common_voice_en_19808301.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6377 + }, + { + "path": "common_voice_en_20606867.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6378 + }, + { + "path": "common_voice_en_35697775.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6379 + }, + { + "path": "common_voice_en_19415544.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6380 + }, + { + "path": "common_voice_en_21730661.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6381 + }, + { + "path": "common_voice_en_19149076.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6382 + }, + { + "path": "common_voice_en_17912300.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6383 + }, + { + "path": "common_voice_en_17269125.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6384 + }, + { + "path": "common_voice_en_19939049.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6385 + }, + { + "path": "common_voice_en_19936427.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6386 + }, + { + "path": "common_voice_en_27734891.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6387 + }, + { + "path": "common_voice_en_31500153.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6388 + }, + { + "path": "common_voice_en_30477718.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6389 + }, + { + "path": "common_voice_en_18262069.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6390 + }, + { + "path": "common_voice_en_18257271.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6391 + }, + { + "path": "common_voice_en_34897397.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6392 + }, + { + "path": "common_voice_en_552057.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6393 + }, + { + "path": "common_voice_en_26746241.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6394 + }, + { + "path": "common_voice_en_17251659.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6395 + }, + { + "path": "common_voice_en_19752014.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6396 + }, + { + "path": "common_voice_en_18345208.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6397 + }, + { + "path": "common_voice_en_31798556.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6398 + }, + { + "path": "common_voice_en_31610538.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6399 + }, + { + "path": "common_voice_en_22544247.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6400 + }, + { + "path": "common_voice_en_18278070.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6401 + }, + { + "path": "common_voice_en_19823532.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6402 + }, + { + "path": "common_voice_en_35437451.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6403 + }, + { + "path": "common_voice_en_33550583.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6404 + }, + { + "path": "common_voice_en_19686592.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6405 + }, + { + "path": "common_voice_en_17250062.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6406 + }, + { + "path": "common_voice_en_34980360.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6407 + }, + { + "path": "common_voice_en_19650760.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6408 + }, + { + "path": "common_voice_en_25256895.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6409 + }, + { + "path": "common_voice_en_22588894.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6410 + }, + { + "path": "common_voice_en_37236607.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6411 + }, + { + "path": "common_voice_en_18710315.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6412 + }, + { + "path": "common_voice_en_648457.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6413 + }, + { + "path": "common_voice_en_335848.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6414 + }, + { + "path": "common_voice_en_20526312.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6415 + }, + { + "path": "common_voice_en_18508472.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6416 + }, + { + "path": "common_voice_en_31713069.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6417 + }, + { + "path": "common_voice_en_19412382.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6418 + }, + { + "path": "common_voice_en_13900.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6419 + }, + { + "path": "common_voice_en_20137222.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6420 + }, + { + "path": "common_voice_en_18777448.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6421 + }, + { + "path": "common_voice_en_32794984.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6422 + }, + { + "path": "common_voice_en_26974255.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6423 + }, + { + "path": "common_voice_en_17782101.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6424 + }, + { + "path": "common_voice_en_22652665.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6425 + }, + { + "path": "common_voice_en_17336131.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6426 + }, + { + "path": "common_voice_en_22334090.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6427 + }, + { + "path": "common_voice_en_32645931.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6428 + }, + { + "path": "common_voice_en_19937020.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6429 + }, + { + "path": "common_voice_en_20976961.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6430 + }, + { + "path": "common_voice_en_36524012.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6431 + }, + { + "path": "common_voice_en_17819209.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6432 + }, + { + "path": "common_voice_en_577935.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6433 + }, + { + "path": "common_voice_en_19703287.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6434 + }, + { + "path": "common_voice_en_19725499.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6435 + }, + { + "path": "common_voice_en_31822774.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6436 + }, + { + "path": "common_voice_en_20540544.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6437 + }, + { + "path": "common_voice_en_19945306.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6438 + }, + { + "path": "common_voice_en_25423290.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6439 + }, + { + "path": "common_voice_en_17288918.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6440 + }, + { + "path": "common_voice_en_17877327.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6441 + }, + { + "path": "common_voice_en_18070362.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6442 + }, + { + "path": "common_voice_en_36290075.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6443 + }, + { + "path": "common_voice_en_27076011.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6444 + }, + { + "path": "common_voice_en_19774282.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6445 + }, + { + "path": "common_voice_en_19757703.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6446 + }, + { + "path": "common_voice_en_18408619.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6447 + }, + { + "path": "common_voice_en_21092395.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6448 + }, + { + "path": "common_voice_en_31671570.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6449 + }, + { + "path": "common_voice_en_64837.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6450 + }, + { + "path": "common_voice_en_606948.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6451 + }, + { + "path": "common_voice_en_32351071.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6452 + }, + { + "path": "common_voice_en_19485661.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6453 + }, + { + "path": "common_voice_en_31573766.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6454 + }, + { + "path": "common_voice_en_23805514.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6455 + }, + { + "path": "common_voice_en_31680894.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6456 + }, + { + "path": "common_voice_en_27095220.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6457 + }, + { + "path": "common_voice_en_19735704.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6458 + }, + { + "path": "common_voice_en_3900314.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6459 + }, + { + "path": "common_voice_en_19708695.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6460 + }, + { + "path": "common_voice_en_20451079.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6461 + }, + { + "path": "common_voice_en_25465455.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6462 + }, + { + "path": "common_voice_en_25635320.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6463 + }, + { + "path": "common_voice_en_18710312.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6464 + }, + { + "path": "common_voice_en_20278402.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6465 + }, + { + "path": "common_voice_en_32960342.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6466 + }, + { + "path": "common_voice_en_19961022.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6467 + }, + { + "path": "common_voice_en_126409.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6468 + }, + { + "path": "common_voice_en_20282437.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6469 + }, + { + "path": "common_voice_en_572305.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6470 + }, + { + "path": "common_voice_en_19941679.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6471 + }, + { + "path": "common_voice_en_33136025.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6472 + }, + { + "path": "common_voice_en_21568053.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6473 + }, + { + "path": "common_voice_en_18707935.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6474 + }, + { + "path": "common_voice_en_19480240.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6475 + }, + { + "path": "common_voice_en_23964609.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6476 + }, + { + "path": "common_voice_en_19373305.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6477 + }, + { + "path": "common_voice_en_37133946.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6478 + }, + { + "path": "common_voice_en_29402881.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6479 + }, + { + "path": "common_voice_en_31661602.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6480 + }, + { + "path": "common_voice_en_23846481.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6481 + }, + { + "path": "common_voice_en_18844016.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6482 + }, + { + "path": "common_voice_en_36886328.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6483 + }, + { + "path": "common_voice_en_21482.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6484 + }, + { + "path": "common_voice_en_36522655.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6485 + }, + { + "path": "common_voice_en_19954083.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6486 + }, + { + "path": "common_voice_en_31578412.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6487 + }, + { + "path": "common_voice_en_18669344.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6488 + }, + { + "path": "common_voice_en_33193617.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6489 + }, + { + "path": "common_voice_en_21573426.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6490 + }, + { + "path": "common_voice_en_18813645.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6491 + }, + { + "path": "common_voice_en_18737952.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6492 + }, + { + "path": "common_voice_en_30615188.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6493 + }, + { + "path": "common_voice_en_27619474.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6494 + }, + { + "path": "common_voice_en_20115558.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6495 + }, + { + "path": "common_voice_en_187061.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6496 + }, + { + "path": "common_voice_en_20688689.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6497 + }, + { + "path": "common_voice_en_23813868.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6498 + }, + { + "path": "common_voice_en_20907101.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6499 + }, + { + "path": "common_voice_en_18041350.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6500 + }, + { + "path": "common_voice_en_37266634.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6501 + }, + { + "path": "common_voice_en_29540105.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6502 + }, + { + "path": "common_voice_en_18276812.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6503 + }, + { + "path": "common_voice_en_20889724.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6504 + }, + { + "path": "common_voice_en_32653578.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6505 + }, + { + "path": "common_voice_en_531902.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6506 + }, + { + "path": "common_voice_en_20278400.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6507 + }, + { + "path": "common_voice_en_26984718.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6508 + }, + { + "path": "common_voice_en_33000196.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6509 + }, + { + "path": "common_voice_en_17904824.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6510 + }, + { + "path": "common_voice_en_1075.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6511 + }, + { + "path": "common_voice_en_123648.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6512 + }, + { + "path": "common_voice_en_31599503.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6513 + }, + { + "path": "common_voice_en_19606474.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6514 + }, + { + "path": "common_voice_en_31728110.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6515 + }, + { + "path": "common_voice_en_30471104.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6516 + }, + { + "path": "common_voice_en_20298793.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6517 + }, + { + "path": "common_voice_en_21230886.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6518 + }, + { + "path": "common_voice_en_17810374.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6519 + }, + { + "path": "common_voice_en_691448.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6520 + }, + { + "path": "common_voice_en_18273546.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6521 + }, + { + "path": "common_voice_en_17357047.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6522 + }, + { + "path": "common_voice_en_36835918.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6523 + }, + { + "path": "common_voice_en_20406861.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6524 + }, + { + "path": "common_voice_en_10410522.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6525 + }, + { + "path": "common_voice_en_25840871.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6526 + }, + { + "path": "common_voice_en_31681671.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6527 + }, + { + "path": "common_voice_en_495881.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6528 + }, + { + "path": "common_voice_en_21284877.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6529 + }, + { + "path": "common_voice_en_495879.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6530 + }, + { + "path": "common_voice_en_21284881.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6531 + }, + { + "path": "common_voice_en_31718095.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6532 + }, + { + "path": "common_voice_en_20848164.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6533 + }, + { + "path": "common_voice_en_20203640.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6534 + }, + { + "path": "common_voice_en_34954464.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6535 + }, + { + "path": "common_voice_en_31538690.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6536 + }, + { + "path": "common_voice_en_32647424.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6537 + }, + { + "path": "common_voice_en_18267179.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6538 + }, + { + "path": "common_voice_en_20221791.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6539 + }, + { + "path": "common_voice_en_19648410.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6540 + }, + { + "path": "common_voice_en_16047346.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6541 + }, + { + "path": "common_voice_en_109017.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6542 + }, + { + "path": "common_voice_en_20304587.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6543 + }, + { + "path": "common_voice_en_30699225.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6544 + }, + { + "path": "common_voice_en_18218583.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6545 + }, + { + "path": "common_voice_en_31563876.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6546 + }, + { + "path": "common_voice_en_19659450.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6547 + }, + { + "path": "common_voice_en_18553336.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6548 + }, + { + "path": "common_voice_en_32233084.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6549 + }, + { + "path": "common_voice_en_33017669.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6550 + }, + { + "path": "common_voice_en_20542865.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6551 + }, + { + "path": "common_voice_en_20984231.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6552 + }, + { + "path": "common_voice_en_19125679.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6553 + }, + { + "path": "common_voice_en_27385248.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6554 + }, + { + "path": "common_voice_en_129916.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6555 + }, + { + "path": "common_voice_en_17261384.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6556 + }, + { + "path": "common_voice_en_18553335.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6557 + }, + { + "path": "common_voice_en_20400528.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6558 + }, + { + "path": "common_voice_en_31772043.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6559 + }, + { + "path": "common_voice_en_20548238.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6560 + }, + { + "path": "common_voice_en_693986.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6561 + }, + { + "path": "common_voice_en_20489056.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6562 + }, + { + "path": "common_voice_en_19988706.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6563 + }, + { + "path": "common_voice_en_28889777.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6564 + }, + { + "path": "common_voice_en_18712894.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6565 + }, + { + "path": "common_voice_en_30389809.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6566 + }, + { + "path": "common_voice_en_21367755.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6567 + }, + { + "path": "common_voice_en_17271034.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6568 + }, + { + "path": "common_voice_en_18669343.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6569 + }, + { + "path": "common_voice_en_193485.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6570 + }, + { + "path": "common_voice_en_37159307.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6571 + }, + { + "path": "common_voice_en_18723995.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6572 + }, + { + "path": "common_voice_en_31641526.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6573 + }, + { + "path": "common_voice_en_27697437.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6574 + }, + { + "path": "common_voice_en_23593451.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6575 + }, + { + "path": "common_voice_en_18165191.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6576 + }, + { + "path": "common_voice_en_32941670.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6577 + }, + { + "path": "common_voice_en_26790804.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6578 + }, + { + "path": "common_voice_en_18165213.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6579 + }, + { + "path": "common_voice_en_673730.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6580 + }, + { + "path": "common_voice_en_20002083.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6581 + }, + { + "path": "common_voice_en_86581.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6582 + }, + { + "path": "common_voice_en_17914760.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6583 + }, + { + "path": "common_voice_en_33843676.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6584 + }, + { + "path": "common_voice_en_18758345.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6585 + }, + { + "path": "common_voice_en_17789645.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6586 + }, + { + "path": "common_voice_en_32719194.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6587 + }, + { + "path": "common_voice_en_22959122.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6588 + }, + { + "path": "common_voice_en_20928833.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6589 + }, + { + "path": "common_voice_en_19932660.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6590 + }, + { + "path": "common_voice_en_16692475.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6591 + }, + { + "path": "common_voice_en_26072506.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6592 + }, + { + "path": "common_voice_en_26950526.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6593 + }, + { + "path": "common_voice_en_32650354.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6594 + }, + { + "path": "common_voice_en_21572998.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6595 + }, + { + "path": "common_voice_en_35016253.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6596 + }, + { + "path": "common_voice_en_17257749.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6597 + }, + { + "path": "common_voice_en_17846037.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6598 + }, + { + "path": "common_voice_en_22790803.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6599 + }, + { + "path": "common_voice_en_32351068.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6600 + }, + { + "path": "common_voice_en_19708705.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6601 + }, + { + "path": "common_voice_en_18316463.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6602 + }, + { + "path": "common_voice_en_19725350.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6603 + }, + { + "path": "common_voice_en_26351936.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6604 + }, + { + "path": "common_voice_en_21157123.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6605 + }, + { + "path": "common_voice_en_22727683.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6606 + }, + { + "path": "common_voice_en_30407409.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6607 + }, + { + "path": "common_voice_en_17359854.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6608 + }, + { + "path": "common_voice_en_18198686.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6609 + }, + { + "path": "common_voice_en_17562784.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6610 + }, + { + "path": "common_voice_en_19611935.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6611 + }, + { + "path": "common_voice_en_30585701.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6612 + }, + { + "path": "common_voice_en_534194.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6613 + }, + { + "path": "common_voice_en_20027504.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6614 + }, + { + "path": "common_voice_en_23756204.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6615 + }, + { + "path": "common_voice_en_17732764.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6616 + }, + { + "path": "common_voice_en_31536330.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6617 + }, + { + "path": "common_voice_en_18274245.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6618 + }, + { + "path": "common_voice_en_19935033.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6619 + }, + { + "path": "common_voice_en_22730454.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6620 + }, + { + "path": "common_voice_en_60360.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6621 + }, + { + "path": "common_voice_en_19988809.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6622 + }, + { + "path": "common_voice_en_20331229.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6623 + }, + { + "path": "common_voice_en_9684911.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6624 + }, + { + "path": "common_voice_en_25650380.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6625 + }, + { + "path": "common_voice_en_17257751.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6626 + }, + { + "path": "common_voice_en_18514959.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6627 + }, + { + "path": "common_voice_en_32960341.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6628 + }, + { + "path": "common_voice_en_22926410.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6629 + }, + { + "path": "common_voice_en_32092487.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6630 + }, + { + "path": "common_voice_en_20000751.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6631 + }, + { + "path": "common_voice_en_20202676.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6632 + }, + { + "path": "common_voice_en_32377446.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6633 + }, + { + "path": "common_voice_en_553505.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6634 + }, + { + "path": "common_voice_en_18855071.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6635 + }, + { + "path": "common_voice_en_30560269.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6636 + }, + { + "path": "common_voice_en_23379807.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6637 + }, + { + "path": "common_voice_en_18255775.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6638 + }, + { + "path": "common_voice_en_23747204.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6639 + }, + { + "path": "common_voice_en_31874234.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6640 + }, + { + "path": "common_voice_en_31757413.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6641 + }, + { + "path": "common_voice_en_36398479.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6642 + }, + { + "path": "common_voice_en_31007975.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6643 + }, + { + "path": "common_voice_en_31609934.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6644 + }, + { + "path": "common_voice_en_18754523.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6645 + }, + { + "path": "common_voice_en_18856235.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6646 + }, + { + "path": "common_voice_en_18170498.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6647 + }, + { + "path": "common_voice_en_30321989.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6648 + }, + { + "path": "common_voice_en_17864433.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6649 + }, + { + "path": "common_voice_en_636862.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6650 + }, + { + "path": "common_voice_en_19774281.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6651 + }, + { + "path": "common_voice_en_33471260.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6652 + }, + { + "path": "common_voice_en_19617112.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6653 + }, + { + "path": "common_voice_en_31752832.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6654 + }, + { + "path": "common_voice_en_37172216.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6655 + }, + { + "path": "common_voice_en_30713098.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6656 + }, + { + "path": "common_voice_en_20688692.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6657 + }, + { + "path": "common_voice_en_20937832.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6658 + }, + { + "path": "common_voice_en_30321993.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6659 + }, + { + "path": "common_voice_en_18712692.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6660 + }, + { + "path": "common_voice_en_600880.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6661 + }, + { + "path": "common_voice_en_25224986.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6662 + }, + { + "path": "common_voice_en_26841931.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6663 + }, + { + "path": "common_voice_en_31582593.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6664 + }, + { + "path": "common_voice_en_31893853.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6665 + }, + { + "path": "common_voice_en_17782099.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6666 + }, + { + "path": "common_voice_en_19489428.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6667 + }, + { + "path": "common_voice_en_31665419.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6668 + }, + { + "path": "common_voice_en_18547963.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6669 + }, + { + "path": "common_voice_en_20273153.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6670 + }, + { + "path": "common_voice_en_35920080.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6671 + }, + { + "path": "common_voice_en_36027911.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6672 + }, + { + "path": "common_voice_en_20895162.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6673 + }, + { + "path": "common_voice_en_20009024.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6674 + }, + { + "path": "common_voice_en_17882304.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6675 + }, + { + "path": "common_voice_en_31548448.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6676 + }, + { + "path": "common_voice_en_20586173.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6677 + }, + { + "path": "common_voice_en_18984784.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6678 + }, + { + "path": "common_voice_en_32293894.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6679 + }, + { + "path": "common_voice_en_27067974.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6680 + }, + { + "path": "common_voice_en_463886.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6681 + }, + { + "path": "common_voice_en_19966755.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6682 + }, + { + "path": "common_voice_en_17669825.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6683 + }, + { + "path": "common_voice_en_36528130.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6684 + }, + { + "path": "common_voice_en_31582705.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6685 + }, + { + "path": "common_voice_en_20688629.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6686 + }, + { + "path": "common_voice_en_33321330.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6687 + }, + { + "path": "common_voice_en_33280385.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6688 + }, + { + "path": "common_voice_en_21780812.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6689 + }, + { + "path": "common_voice_en_34989985.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6690 + }, + { + "path": "common_voice_en_18733679.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6691 + }, + { + "path": "common_voice_en_20234122.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6692 + }, + { + "path": "common_voice_en_18355954.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6693 + }, + { + "path": "common_voice_en_20259678.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6694 + }, + { + "path": "common_voice_en_17250888.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6695 + }, + { + "path": "common_voice_en_678752.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6696 + }, + { + "path": "common_voice_en_19932747.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6697 + }, + { + "path": "common_voice_en_18945361.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6698 + }, + { + "path": "common_voice_en_21372561.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6699 + }, + { + "path": "common_voice_en_20648397.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6700 + }, + { + "path": "common_voice_en_37237056.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6701 + }, + { + "path": "common_voice_en_21369377.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6702 + }, + { + "path": "common_voice_en_34954466.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6703 + }, + { + "path": "common_voice_en_19442319.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6704 + }, + { + "path": "common_voice_en_18352995.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6705 + }, + { + "path": "common_voice_en_18355955.mp3", + "question": "As per the speech's vocal sound, what age group seems fitting?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6706 + }, + { + "path": "common_voice_en_596447.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6707 + }, + { + "path": "common_voice_en_17760442.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6708 + }, + { + "path": "common_voice_en_20282120.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6709 + }, + { + "path": "common_voice_en_8805512.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6710 + }, + { + "path": "common_voice_en_36376466.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6711 + }, + { + "path": "common_voice_en_20242381.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6712 + }, + { + "path": "common_voice_en_19050518.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6713 + }, + { + "path": "common_voice_en_20930829.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6714 + }, + { + "path": "common_voice_en_13483770.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6715 + }, + { + "path": "common_voice_en_20016035.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6716 + }, + { + "path": "common_voice_en_17786123.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6717 + }, + { + "path": "common_voice_en_34571825.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6718 + }, + { + "path": "common_voice_en_32321230.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6719 + }, + { + "path": "common_voice_en_28197049.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6720 + }, + { + "path": "common_voice_en_5688714.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6721 + }, + { + "path": "common_voice_en_560792.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6722 + }, + { + "path": "common_voice_en_17562715.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6723 + }, + { + "path": "common_voice_en_31673450.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6724 + }, + { + "path": "common_voice_en_26357.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6725 + }, + { + "path": "common_voice_en_19945342.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6726 + }, + { + "path": "common_voice_en_19539287.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6727 + }, + { + "path": "common_voice_en_19647545.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6728 + }, + { + "path": "common_voice_en_22754640.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6729 + }, + { + "path": "common_voice_en_23956394.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6730 + }, + { + "path": "common_voice_en_19539213.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6731 + }, + { + "path": "common_voice_en_20192653.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6732 + }, + { + "path": "common_voice_en_934751.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6733 + }, + { + "path": "common_voice_en_31788782.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6734 + }, + { + "path": "common_voice_en_18661019.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6735 + }, + { + "path": "common_voice_en_19927944.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6736 + }, + { + "path": "common_voice_en_37016367.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6737 + }, + { + "path": "common_voice_en_18758956.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6738 + }, + { + "path": "common_voice_en_25659718.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6739 + }, + { + "path": "common_voice_en_19464263.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6740 + }, + { + "path": "common_voice_en_23774819.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6741 + }, + { + "path": "common_voice_en_18854895.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6742 + }, + { + "path": "common_voice_en_554398.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6743 + }, + { + "path": "common_voice_en_36938662.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6744 + }, + { + "path": "common_voice_en_19725497.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6745 + }, + { + "path": "common_voice_en_19751839.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6746 + }, + { + "path": "common_voice_en_31618142.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6747 + }, + { + "path": "common_voice_en_18012942.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6748 + }, + { + "path": "common_voice_en_29248868.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6749 + }, + { + "path": "common_voice_en_26233296.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6750 + }, + { + "path": "common_voice_en_19700501.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6751 + }, + { + "path": "common_voice_en_18849782.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6752 + }, + { + "path": "common_voice_en_19706318.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6753 + }, + { + "path": "common_voice_en_25705539.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6754 + }, + { + "path": "common_voice_en_19757182.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6755 + }, + { + "path": "common_voice_en_35092634.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6756 + }, + { + "path": "common_voice_en_495880.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6757 + }, + { + "path": "common_voice_en_17291039.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6758 + }, + { + "path": "common_voice_en_30535397.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6759 + }, + { + "path": "common_voice_en_24036718.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6760 + }, + { + "path": "common_voice_en_31070493.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6761 + }, + { + "path": "common_voice_en_18306537.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6762 + }, + { + "path": "common_voice_en_37159498.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6763 + }, + { + "path": "common_voice_en_18062241.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6764 + }, + { + "path": "common_voice_en_19666442.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6765 + }, + { + "path": "common_voice_en_18955915.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6766 + }, + { + "path": "common_voice_en_31713475.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6767 + }, + { + "path": "common_voice_en_18295850.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6768 + }, + { + "path": "common_voice_en_22469604.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6769 + }, + { + "path": "common_voice_en_10091129.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6770 + }, + { + "path": "common_voice_en_698681.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6771 + }, + { + "path": "common_voice_en_36737882.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6772 + }, + { + "path": "common_voice_en_18723909.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6773 + }, + { + "path": "common_voice_en_512959.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6774 + }, + { + "path": "common_voice_en_636859.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6775 + }, + { + "path": "common_voice_en_19609644.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6776 + }, + { + "path": "common_voice_en_19816038.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6777 + }, + { + "path": "common_voice_en_19985822.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6778 + }, + { + "path": "common_voice_en_34842726.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6779 + }, + { + "path": "common_voice_en_32118312.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6780 + }, + { + "path": "common_voice_en_21307080.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6781 + }, + { + "path": "common_voice_en_19311994.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6782 + }, + { + "path": "common_voice_en_19742276.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6783 + }, + { + "path": "common_voice_en_30981013.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6784 + }, + { + "path": "common_voice_en_21157121.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6785 + }, + { + "path": "common_voice_en_22575527.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6786 + }, + { + "path": "common_voice_en_17761046.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6787 + }, + { + "path": "common_voice_en_20825528.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6788 + }, + { + "path": "common_voice_en_31704224.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6789 + }, + { + "path": "common_voice_en_31522602.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6790 + }, + { + "path": "common_voice_en_37016755.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6791 + }, + { + "path": "common_voice_en_18506357.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6792 + }, + { + "path": "common_voice_en_31596100.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6793 + }, + { + "path": "common_voice_en_1323348.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6794 + }, + { + "path": "common_voice_en_19941145.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6795 + }, + { + "path": "common_voice_en_36435650.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6796 + }, + { + "path": "common_voice_en_20374635.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6797 + }, + { + "path": "common_voice_en_30998559.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6798 + }, + { + "path": "common_voice_en_18309499.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6799 + }, + { + "path": "common_voice_en_26408640.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6800 + }, + { + "path": "common_voice_en_21332805.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6801 + }, + { + "path": "common_voice_en_23935850.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6802 + }, + { + "path": "common_voice_en_23628350.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6803 + }, + { + "path": "common_voice_en_671188.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6804 + }, + { + "path": "common_voice_en_19270939.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6805 + }, + { + "path": "common_voice_en_18100938.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6806 + }, + { + "path": "common_voice_en_17285930.mp3", + "question": "From the speaker's elocution, can you deduce an age range?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6807 + }, + { + "path": "common_voice_en_23747206.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6808 + }, + { + "path": "common_voice_en_36003968.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6809 + }, + { + "path": "common_voice_en_22709018.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6810 + }, + { + "path": "common_voice_en_17934372.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6811 + }, + { + "path": "common_voice_en_20644782.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6812 + }, + { + "path": "common_voice_en_23583679.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6813 + }, + { + "path": "common_voice_en_20928834.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6814 + }, + { + "path": "common_voice_en_138772.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6815 + }, + { + "path": "common_voice_en_36939175.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6816 + }, + { + "path": "common_voice_en_20426389.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6817 + }, + { + "path": "common_voice_en_19532961.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6818 + }, + { + "path": "common_voice_en_18844018.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6819 + }, + { + "path": "common_voice_en_18242955.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6820 + }, + { + "path": "common_voice_en_20301324.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6821 + }, + { + "path": "common_voice_en_31685260.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6822 + }, + { + "path": "common_voice_en_21685241.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6823 + }, + { + "path": "common_voice_en_35714312.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6824 + }, + { + "path": "common_voice_en_37220997.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6825 + }, + { + "path": "common_voice_en_18554229.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6826 + }, + { + "path": "common_voice_en_27102746.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6827 + }, + { + "path": "common_voice_en_19213279.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6828 + }, + { + "path": "common_voice_en_17723402.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6829 + }, + { + "path": "common_voice_en_19141245.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6830 + }, + { + "path": "common_voice_en_17865774.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6831 + }, + { + "path": "common_voice_en_31798558.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6832 + }, + { + "path": "common_voice_en_31661600.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6833 + }, + { + "path": "common_voice_en_17291041.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6834 + }, + { + "path": "common_voice_en_22967210.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6835 + }, + { + "path": "common_voice_en_19941526.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6836 + }, + { + "path": "common_voice_en_23896971.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6837 + }, + { + "path": "common_voice_en_20296863.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6838 + }, + { + "path": "common_voice_en_22910971.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6839 + }, + { + "path": "common_voice_en_368765.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6840 + }, + { + "path": "common_voice_en_19141242.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6841 + }, + { + "path": "common_voice_en_23593449.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6842 + }, + { + "path": "common_voice_en_20400527.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6843 + }, + { + "path": "common_voice_en_31770143.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6844 + }, + { + "path": "common_voice_en_21798355.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6845 + }, + { + "path": "common_voice_en_21154507.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6846 + }, + { + "path": "common_voice_en_20380244.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6847 + }, + { + "path": "common_voice_en_33409250.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6848 + }, + { + "path": "common_voice_en_35379578.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6849 + }, + { + "path": "common_voice_en_85509.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6850 + }, + { + "path": "common_voice_en_31716424.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6851 + }, + { + "path": "common_voice_en_31661956.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6852 + }, + { + "path": "common_voice_en_17294097.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6853 + }, + { + "path": "common_voice_en_19622846.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6854 + }, + { + "path": "common_voice_en_22948690.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6855 + }, + { + "path": "common_voice_en_32538390.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6856 + }, + { + "path": "common_voice_en_23956396.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6857 + }, + { + "path": "common_voice_en_19836714.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6858 + }, + { + "path": "common_voice_en_689423.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6859 + }, + { + "path": "common_voice_en_203202.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6860 + }, + { + "path": "common_voice_en_19654564.mp3", + "question": "What age division does the speaker's voice pitch suggest?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6861 + }, + { + "path": "common_voice_en_34399333.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6862 + }, + { + "path": "common_voice_en_20987283.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6863 + }, + { + "path": "common_voice_en_17365160.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6864 + }, + { + "path": "common_voice_en_19058127.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6865 + }, + { + "path": "common_voice_en_19973696.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6866 + }, + { + "path": "common_voice_en_18856231.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6867 + }, + { + "path": "common_voice_en_27731438.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6868 + }, + { + "path": "common_voice_en_27257379.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6869 + }, + { + "path": "common_voice_en_31695208.mp3", + "question": "In listening to the speech, what age could you attribute to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6870 + }, + { + "path": "common_voice_en_19659449.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6871 + }, + { + "path": "common_voice_en_17787575.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6872 + }, + { + "path": "common_voice_en_25722369.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6873 + }, + { + "path": "common_voice_en_17265071.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6874 + }, + { + "path": "common_voice_en_37230615.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6875 + }, + { + "path": "common_voice_en_31871816.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6876 + }, + { + "path": "common_voice_en_25618450.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6877 + }, + { + "path": "common_voice_en_21374777.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6878 + }, + { + "path": "common_voice_en_31713487.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6879 + }, + { + "path": "common_voice_en_18317514.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6880 + }, + { + "path": "common_voice_en_19212949.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6881 + }, + { + "path": "common_voice_en_18882314.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6882 + }, + { + "path": "common_voice_en_21307079.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6883 + }, + { + "path": "common_voice_en_19932657.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6884 + }, + { + "path": "common_voice_en_22699569.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6885 + }, + { + "path": "common_voice_en_19941683.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6886 + }, + { + "path": "common_voice_en_1272597.mp3", + "question": "To which age demographic does the speaker's speech pattern align?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6887 + }, + { + "path": "common_voice_en_18425397.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6888 + }, + { + "path": "common_voice_en_17414289.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6889 + }, + { + "path": "common_voice_en_18967266.mp3", + "question": "In your opinion, what age category does the speaker's voice fit into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6890 + }, + { + "path": "common_voice_en_20502347.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6891 + }, + { + "path": "common_voice_en_23628352.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6892 + }, + { + "path": "common_voice_en_23768959.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6893 + }, + { + "path": "common_voice_en_18577449.mp3", + "question": "From the tone of the speech, how old is the speaker likely to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6894 + }, + { + "path": "common_voice_en_1175989.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6895 + }, + { + "path": "common_voice_en_573398.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6896 + }, + { + "path": "common_voice_en_17266452.mp3", + "question": "What age group does the speaker's voice seem to indicate?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6897 + }, + { + "path": "common_voice_en_37040119.mp3", + "question": "Which of the given age groups does the speaker's voice resemble?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6898 + }, + { + "path": "common_voice_en_19933879.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6899 + }, + { + "path": "common_voice_en_17281445.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6900 + }, + { + "path": "common_voice_en_516191.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6901 + }, + { + "path": "common_voice_en_35324672.mp3", + "question": "What age does the speaker's speech suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6902 + }, + { + "path": "common_voice_en_2127639.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6903 + }, + { + "path": "common_voice_en_623389.mp3", + "question": "How would you gauge the speaker's age based on their voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6904 + }, + { + "path": "common_voice_en_28847861.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6905 + }, + { + "path": "common_voice_en_31584058.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6906 + }, + { + "path": "common_voice_en_20048164.mp3", + "question": "What is your best guess for the speaker's age category?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6907 + }, + { + "path": "common_voice_en_21943181.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6908 + }, + { + "path": "common_voice_en_20338892.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6909 + }, + { + "path": "common_voice_en_20586174.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6910 + }, + { + "path": "common_voice_en_31206949.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6911 + }, + { + "path": "common_voice_en_15734835.mp3", + "question": "What age group would you say the speaker's voice belongs to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6912 + }, + { + "path": "common_voice_en_20025414.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6913 + }, + { + "path": "common_voice_en_17298948.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6914 + }, + { + "path": "common_voice_en_22469599.mp3", + "question": "Which age range do you believe best matches the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6915 + }, + { + "path": "common_voice_en_31516693.mp3", + "question": "How would you profile the speaker's age based on their vocal sound?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6916 + }, + { + "path": "common_voice_en_22713796.mp3", + "question": "What age range does the speaker's tone suggest to you?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6917 + }, + { + "path": "common_voice_en_25619619.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6918 + }, + { + "path": "common_voice_en_64648.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6919 + }, + { + "path": "common_voice_en_17492946.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6920 + }, + { + "path": "common_voice_en_31512428.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6921 + }, + { + "path": "common_voice_en_18545940.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6922 + }, + { + "path": "common_voice_en_20009192.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6923 + }, + { + "path": "common_voice_en_17780603.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6924 + }, + { + "path": "common_voice_en_20027505.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6925 + }, + { + "path": "common_voice_en_34917235.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6926 + }, + { + "path": "common_voice_en_27069841.mp3", + "question": "Considering the clarity of the speech, what age might the speaker represent?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6927 + }, + { + "path": "common_voice_en_30633200.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6928 + }, + { + "path": "common_voice_en_23116353.mp3", + "question": "From the speaker's articulation, what age could they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6929 + }, + { + "path": "common_voice_en_20937834.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6930 + }, + { + "path": "common_voice_en_34082125.mp3", + "question": "From the speaker's speaking style, what age bracket is indicated?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6931 + }, + { + "path": "common_voice_en_17252412.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6932 + }, + { + "path": "common_voice_en_18260627.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6933 + }, + { + "path": "common_voice_en_595872.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "fifties to sixties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6934 + }, + { + "path": "common_voice_en_31615503.mp3", + "question": "What age bracket is conveyed through the speaker's vocal expression?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6935 + }, + { + "path": "common_voice_en_329504.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6936 + }, + { + "path": "common_voice_en_19545133.mp3", + "question": "What age seems to be represented by the speaker's vocal pitch?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6937 + }, + { + "path": "common_voice_en_30471101.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6938 + }, + { + "path": "common_voice_en_31703178.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6939 + }, + { + "path": "common_voice_en_19643569.mp3", + "question": "How old do you think the speaker is based on their vocal characteristics?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6940 + }, + { + "path": "common_voice_en_31704096.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6941 + }, + { + "path": "common_voice_en_17761313.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6942 + }, + { + "path": "common_voice_en_19947345.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6943 + }, + { + "path": "common_voice_en_34891885.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6944 + }, + { + "path": "common_voice_en_37083045.mp3", + "question": "Which age interval does the speaker's voice quality hint at?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6945 + }, + { + "path": "common_voice_en_19708707.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6946 + }, + { + "path": "common_voice_en_20291235.mp3", + "question": "What probable age range can you infer from the speaker's diction?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6947 + }, + { + "path": "common_voice_en_30506189.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6948 + }, + { + "path": "common_voice_en_18137152.mp3", + "question": "Given the speaker's vocal nuances, which age range fits?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6949 + }, + { + "path": "common_voice_en_19396205.mp3", + "question": "Considering the voice, what age band would you ascribe to the speaker?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6950 + }, + { + "path": "common_voice_en_19933256.mp3", + "question": "Based on voice alone, where would you place the speaker's age?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6951 + }, + { + "path": "common_voice_en_21667418.mp3", + "question": "What age spectrum does the speaker's vocal energy reflect?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6952 + }, + { + "path": "common_voice_en_19451160.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6953 + }, + { + "path": "common_voice_en_474450.mp3", + "question": "What age cohort does the speaker's intonation align with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6954 + }, + { + "path": "common_voice_en_18343224.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6955 + }, + { + "path": "common_voice_en_21092392.mp3", + "question": "What is the likely age bracket for the speaker based on their vocal delivery?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6956 + }, + { + "path": "common_voice_en_9684915.mp3", + "question": "To what age range does the speaker's voice seem to correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6957 + }, + { + "path": "common_voice_en_10564.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6958 + }, + { + "path": "common_voice_en_19936681.mp3", + "question": "What is your perception of the speaker's age given their speech quality?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6959 + }, + { + "path": "common_voice_en_18509135.mp3", + "question": "To what age group does the speaker's vocal maturity correspond?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6960 + }, + { + "path": "common_voice_en_30663925.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6961 + }, + { + "path": "common_voice_en_31693756.mp3", + "question": "What is your impression of the speaker's age from their speech tones?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6962 + }, + { + "path": "common_voice_en_17272412.mp3", + "question": "In which age bracket would you classify the speaker based on their speech?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6963 + }, + { + "path": "common_voice_en_19058131.mp3", + "question": "Based on the speech, what would you estimate the speaker's age to be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6964 + }, + { + "path": "common_voice_en_18317544.mp3", + "question": "Which age group does the speaker's voice best correspond with?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6965 + }, + { + "path": "common_voice_en_540648.mp3", + "question": "Based on the speaker's voice timbre, what age might they be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6966 + }, + { + "path": "common_voice_en_17261386.mp3", + "question": "Which age classification seems appropriate for the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6967 + }, + { + "path": "common_voice_en_19854770.mp3", + "question": "What age group does the speaker's voice most likely fall into?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6968 + }, + { + "path": "common_voice_en_22717229.mp3", + "question": "identify the speaker's voice as belonging to a specific age group.", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6969 + }, + { + "path": "common_voice_en_35011971.mp3", + "question": "What age does the speaker's vocal delivery point to?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "teens to twenties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6970 + }, + { + "path": "common_voice_en_26358.mp3", + "question": "Which age group's characteristics are evident in the speaker's voice?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "thirties to fourties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6971 + }, + { + "path": "common_voice_en_32701721.mp3", + "question": "Based on the energy in the speech, what age group could the speaker be?", + "choice_a": "teens to twenties", + "choice_b": "thirties to fourties", + "choice_c": "fifties to sixties", + "choice_d": "seventies to eighties", + "answer_gt": "seventies to eighties", + "task_name": "Speaker_Age_Prediction", + "dataset_name": "common_voice_13.0_en", + "uniq_id": 6972 + }, + { + "path": "audio-1501152468.flac", + "question": "tell me the first 'transport_type'-connected word in this audio.", + "choice_a": "list", + "choice_b": "all", + "choice_c": "train", + "choice_d": "taxi", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6973 + }, + { + "path": "audio-1499089039-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'general_frequency'?", + "choice_a": "take", + "choice_b": "every", + "choice_c": "hourly", + "choice_d": "every day", + "answer_gt": "every day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6974 + }, + { + "path": "audio-1501688874-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'artist_name'?", + "choice_a": "justin bieber's", + "choice_b": "next", + "choice_c": "third day", + "choice_d": "slayer", + "answer_gt": "justin bieber's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6975 + }, + { + "path": "audio-1502114385.flac", + "question": "What is the initial 'person'-associated word you hear in this recording?", + "choice_a": "to", + "choice_b": "email", + "choice_c": "margaret", + "choice_d": "dan", + "answer_gt": "margaret", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6976 + }, + { + "path": "audio--1504192261.flac", + "question": "Which word comes first in this audio that's tied to 'player_setting'?", + "choice_a": "stop", + "choice_b": "stop", + "choice_c": "replay", + "choice_d": "skip", + "answer_gt": "stop", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6977 + }, + { + "path": "audio--1506078251.flac", + "question": "tell me the first word that's in relation to 'alarm_type' in this audio?", + "choice_a": "please", + "choice_b": "wake", + "choice_c": "wake up", + "choice_d": "soccer practice", + "answer_gt": "wake up", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6978 + }, + { + "path": "audio-1490199787.flac", + "question": "What's the first word that's tied to 'media_type' mentioned in this audio?", + "choice_a": "tweet", + "choice_b": "complaint", + "choice_c": "facebook account", + "choice_d": "newest provider", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6979 + }, + { + "path": "audio-1495369316-headset.flac", + "question": "What is the first word in this clip that's relevant to 'person'?", + "choice_a": "morgan freeman", + "choice_b": "movie", + "choice_c": "clint eastwood", + "choice_d": "jack", + "answer_gt": "morgan freeman", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6980 + }, + { + "path": "audio-1492784702-headset.flac", + "question": "Which word that's affiliated with 'media_type' is spoken first in this clip?", + "choice_a": "cnn", + "choice_b": "are", + "choice_c": "bob's news", + "choice_d": "time of india", + "answer_gt": "cnn", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6981 + }, + { + "path": "audio-1501756102-headset.flac", + "question": "What word is the first to reference 'personal_info' in this sound clip?", + "choice_a": "this", + "choice_b": "add", + "choice_c": "phone numbers", + "choice_d": "email", + "answer_gt": "email", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6982 + }, + { + "path": "audio-1496265313.flac", + "question": "Which word that's synonymous with 'media_type' appears first in this audio?", + "choice_a": "the", + "choice_b": "from", + "choice_c": "facebook account", + "choice_d": "cnn", + "answer_gt": "cnn", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6983 + }, + { + "path": "audio-1490200888-headset.flac", + "question": "Which word that's affiliated with 'media_type' is spoken first in this clip?", + "choice_a": "the", + "choice_b": "tweet", + "choice_c": "social networks", + "choice_d": "podcasts", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6984 + }, + { + "path": "audio-1488987034.flac", + "question": "Which word that signifies 'event_name' is spoken first in this clip?", + "choice_a": "next", + "choice_b": "dr", + "choice_c": "meeting", + "choice_d": "appointment", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6985 + }, + { + "path": "audio-1501778342-headset.flac", + "question": "What's the first word that's in reference to 'personal_info' in this sound clip?", + "choice_a": "many", + "choice_b": "for", + "choice_c": "email", + "choice_d": "phone numbers", + "answer_gt": "phone numbers", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6986 + }, + { + "path": "audio-1501764841-headset.flac", + "question": "What word relevant to 'device_type' is mentioned at the start of this audio?", + "choice_a": "vacuum cleaner", + "choice_b": "on", + "choice_c": "wemo plug socket", + "choice_d": "robot roomba", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6987 + }, + { + "path": "audio-1497887650.flac", + "question": "What's the first word that's connected to 'date' in this audio?", + "choice_a": "anything", + "choice_b": "today", + "choice_c": "this week", + "choice_d": "weekdays", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6988 + }, + { + "path": "audio-1489498677.flac", + "question": "What word relevant to 'house_place' is mentioned at the start of this audio?", + "choice_a": "the", + "choice_b": "hoover", + "choice_c": "hallway", + "choice_d": "kitchen", + "answer_gt": "hallway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6989 + }, + { + "path": "audio-1498645533.flac", + "question": "Which word comes first in this audio that's tied to 'event_name'?", + "choice_a": "remove", + "choice_b": "from", + "choice_c": "lunch meeting", + "choice_d": "birthday", + "answer_gt": "birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6990 + }, + { + "path": "audio-1434539121-headset.flac", + "question": "Which word that's synonymous with 'event_name' appears first in this audio?", + "choice_a": "more", + "choice_b": "i", + "choice_c": "parade", + "choice_d": "winter solstice", + "answer_gt": "parade", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6991 + }, + { + "path": "audio-1501408361-headset.flac", + "question": "Which word that's affiliated with 'weather_descriptor' is spoken first in this clip?", + "choice_a": "is", + "choice_b": "cold", + "choice_c": "driveway", + "choice_d": "cold", + "answer_gt": "cold", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6992 + }, + { + "path": "audio--1504197602.flac", + "question": "Which word that signifies 'alarm_type' is spoken first in this clip?", + "choice_a": "olly", + "choice_b": "cancel", + "choice_c": "wake up", + "choice_d": "soccer practice", + "answer_gt": "soccer practice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6993 + }, + { + "path": "audio-1497621094-headset.flac", + "question": "tell me the first 'radio_name'-connected word in this audio.", + "choice_a": "start", + "choice_b": "bbc", + "choice_c": "hot one hundred and five", + "choice_d": "bbc", + "answer_gt": "bbc", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6994 + }, + { + "path": "audio-1490192671-headset.flac", + "question": "name the first word that's related to 'transport_type' in this audio?", + "choice_a": "to", + "choice_b": "book", + "choice_c": "taxi", + "choice_d": "cab", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6995 + }, + { + "path": "audio-1490356700-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'relation'?", + "choice_a": "my", + "choice_b": "wife", + "choice_c": "mom", + "choice_d": "boss", + "answer_gt": "wife", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6996 + }, + { + "path": "audio-1500036462.flac", + "question": "name the first word that's related to 'weather_descriptor' in this audio?", + "choice_a": "should", + "choice_b": "jacket", + "choice_c": "rain", + "choice_d": "nice", + "answer_gt": "jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6997 + }, + { + "path": "audio-1497880250.flac", + "question": "detect the first mention of a word related to 'person' in this audio.", + "choice_a": "amy", + "choice_b": "new", + "choice_c": "morgan's", + "choice_d": "petey's", + "answer_gt": "amy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6998 + }, + { + "path": "audio-1490007004-headset.flac", + "question": "What is the first word in this clip that's relevant to 'date'?", + "choice_a": "all", + "choice_b": "one", + "choice_c": "one thousand nine hundred and ninety", + "choice_d": "march eighteen", + "answer_gt": "one thousand nine hundred and ninety", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 6999 + }, + { + "path": "audio-1502218054.flac", + "question": "What's the earliest 'place_name'-related term you hear in this clip?", + "choice_a": "canada", + "choice_b": "current", + "choice_c": "canada", + "choice_d": "library", + "answer_gt": "canada", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7000 + }, + { + "path": "audio-1498569089-headset.flac", + "question": "What's the first word that's tied to 'transport_type' mentioned in this audio?", + "choice_a": "me", + "choice_b": "find", + "choice_c": "metro", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7001 + }, + { + "path": "audio-1501767448.flac", + "question": "In this sound clip, which word related to 'person' is mentioned first?", + "choice_a": "new", + "choice_b": "email", + "choice_c": "ray", + "choice_d": "angeline", + "answer_gt": "ray", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7002 + }, + { + "path": "audio-1490705956.flac", + "question": "What's the first word that's tied to 'person' mentioned in this audio?", + "choice_a": "george", + "choice_b": "that", + "choice_c": "john smith", + "choice_d": "sarah", + "answer_gt": "george", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7003 + }, + { + "path": "audio-1502218273-headset.flac", + "question": "What's the first word that's suggestive of 'place_name' in this audio recording?", + "choice_a": "time", + "choice_b": "and", + "choice_c": "uk", + "choice_d": "world", + "answer_gt": "uk", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7004 + }, + { + "path": "audio-1500980567-headset.flac", + "question": "Which word that's affiliated with 'time' is spoken first in this clip?", + "choice_a": "a quarter to two", + "choice_b": "to", + "choice_c": "before", + "choice_d": "five pm", + "answer_gt": "a quarter to two", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7005 + }, + { + "path": "audio-1499265283.flac", + "question": "What's the first word mentioned in this audio that's about 'place_name'?", + "choice_a": "far", + "choice_b": "canada", + "choice_c": "london", + "choice_d": "california", + "answer_gt": "canada", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7006 + }, + { + "path": "audio-1501758867.flac", + "question": "What's the earliest 'business_name'-related term you hear in this clip?", + "choice_a": "twitter", + "choice_b": "open", + "choice_c": "consumer service", + "choice_d": "i. b. m.", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7007 + }, + { + "path": "audio-1497887581.flac", + "question": "Which word that's affiliated with 'media_type' is spoken first in this clip?", + "choice_a": "complaint", + "choice_b": "online", + "choice_c": "tweet", + "choice_d": "new york times", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7008 + }, + { + "path": "audio-1492780266.flac", + "question": "point out the first 'news_topic'-related word uttered in this audio.", + "choice_a": "news", + "choice_b": "me", + "choice_c": "environmental", + "choice_d": "environment", + "answer_gt": "environment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7009 + }, + { + "path": "audio-1499698164.flac", + "question": "What's the first word that's in reference to 'time' in this sound clip?", + "choice_a": "to", + "choice_b": "a", + "choice_c": "eight am", + "choice_d": "a quarter to two", + "answer_gt": "a quarter to two", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7010 + }, + { + "path": "audio-1497262974.flac", + "question": "name the first word that's related to 'time' in this audio?", + "choice_a": "alarm", + "choice_b": "set", + "choice_c": "six am", + "choice_d": "week", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7011 + }, + { + "path": "audio--1504190990-headset.flac", + "question": "What's the first word that's reflective of 'date' in this sound clip?", + "choice_a": "today", + "choice_b": "alarm", + "choice_c": "a day before", + "choice_d": "thirteenth june", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7012 + }, + { + "path": "audio-1490104820.flac", + "question": "What's the first word that's characteristic of 'place_name' in this clip?", + "choice_a": "work", + "choice_b": "detroit", + "choice_c": "chennai", + "choice_d": "bay area", + "answer_gt": "detroit", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7013 + }, + { + "path": "audio-1501692963-headset.flac", + "question": "Which word linked to 'relation' comes first in this recording?", + "choice_a": "how", + "choice_b": "send", + "choice_c": "friend", + "choice_d": "mom's", + "answer_gt": "friend", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7014 + }, + { + "path": "audio-1497882624-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'list_name'?", + "choice_a": "to do", + "choice_b": "do", + "choice_c": "walmart list", + "choice_d": "groceries", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7015 + }, + { + "path": "audio-1498569735.flac", + "question": "What is the initial 'email_address'-associated word you hear in this recording?", + "choice_a": "dot", + "choice_b": "joseph@gmail dot com", + "choice_c": "macs@gmail dot com", + "choice_d": "john@gmail dot com", + "answer_gt": "joseph@gmail dot com", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7016 + }, + { + "path": "audio-1500900650-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'person'?", + "choice_a": "john", + "choice_b": "please", + "choice_c": "steph curry", + "choice_d": "george eliot's", + "answer_gt": "john", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7017 + }, + { + "path": "audio-1498565025.flac", + "question": "What is the first word in this clip that's relevant to 'weather_descriptor'?", + "choice_a": "the", + "choice_b": "temperature", + "choice_c": "sandals", + "choice_d": "full cover jacket", + "answer_gt": "temperature", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7018 + }, + { + "path": "audio-1498576747.flac", + "question": "What's the earliest 'event_name'-related term you hear in this clip?", + "choice_a": "am", + "choice_b": "meeting", + "choice_c": "haircut", + "choice_d": "wash the windows", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7019 + }, + { + "path": "audio-1501691831-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'place_name'?", + "choice_a": "train station", + "choice_b": "the", + "choice_c": "chennai", + "choice_d": "town", + "answer_gt": "train station", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7020 + }, + { + "path": "audio-1495733082.flac", + "question": "tell me the first 'media_type'-connected word in this audio.", + "choice_a": "tweet", + "choice_b": "at", + "choice_c": "podcasts", + "choice_d": "time of india", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7021 + }, + { + "path": "audio-1502200946.flac", + "question": "What's the first word that's suggestive of 'currency_name' in this audio recording?", + "choice_a": "exchange", + "choice_b": "to", + "choice_c": "dolla", + "choice_d": "dollar", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7022 + }, + { + "path": "audio-1495377638-headset.flac", + "question": "What is the initial 'date'-associated word you hear in this recording?", + "choice_a": "outside", + "choice_b": "today", + "choice_c": "today's", + "choice_d": "todays", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7023 + }, + { + "path": "audio-1492783895-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'time'?", + "choice_a": "three", + "choice_b": "three pm", + "choice_c": "fifteen minutes", + "choice_d": "nine am", + "answer_gt": "three pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7024 + }, + { + "path": "audio-1490286759.flac", + "question": "Which word that's linked to 'time' do you hear first in this recording?", + "choice_a": "thirty minute", + "choice_b": "for", + "choice_c": "twelve pm", + "choice_d": "sunrise", + "answer_gt": "thirty minute", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7025 + }, + { + "path": "audio-1501757591-headset.flac", + "question": "tell me the first word that's in relation to 'person' in this audio?", + "choice_a": "tell", + "choice_b": "billy", + "choice_c": "billy crytals", + "choice_d": "bob", + "answer_gt": "billy crytals", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7026 + }, + { + "path": "audio-1490354834-headset.flac", + "question": "Which term that relates to 'time' is spoken first in this clip?", + "choice_a": "the", + "choice_b": "the", + "choice_c": "midnight", + "choice_d": "rest of the day", + "answer_gt": "rest of the day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7027 + }, + { + "path": "audio-1494420654-headset.flac", + "question": "find the first word that's linked to 'date' in this sound clip?", + "choice_a": "i", + "choice_b": "tomorrow", + "choice_c": "march eighteen", + "choice_d": "wednesday", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7028 + }, + { + "path": "audio-1498570406-headset.flac", + "question": "Which word is the first to be related to 'place_name' in this recording?", + "choice_a": "bolshoi theatre", + "choice_b": "theatre", + "choice_c": "japan", + "choice_d": "alaska", + "answer_gt": "bolshoi theatre", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7029 + }, + { + "path": "audio-1434533748-headset.flac", + "question": "Which word comes first in this audio that's tied to 'list_name'?", + "choice_a": "activities", + "choice_b": "list", + "choice_c": "grocery", + "choice_d": "to do's", + "answer_gt": "activities", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7030 + }, + { + "path": "audio-1499690023-headset.flac", + "question": "point out the first 'event_name'-related word uttered in this audio.", + "choice_a": "does", + "choice_b": "what", + "choice_c": "birthday", + "choice_d": "fringe", + "answer_gt": "birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7031 + }, + { + "path": "audio-1501777713-headset.flac", + "question": "What's the first word that's suggestive of 'place_name' in this audio recording?", + "choice_a": "is", + "choice_b": "time", + "choice_c": "london", + "choice_d": "sydney", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7032 + }, + { + "path": "audio-1501767928-headset.flac", + "question": "Which word comes first in this audio that's tied to 'event_name'?", + "choice_a": "me", + "choice_b": "to", + "choice_c": "wash the windows", + "choice_d": "sports game", + "answer_gt": "wash the windows", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7033 + }, + { + "path": "audio-1502300328.flac", + "question": "What's the first word that's characteristic of 'order_type' in this clip?", + "choice_a": "delivery", + "choice_b": "they", + "choice_c": "takeaway", + "choice_d": "deliver", + "answer_gt": "delivery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7034 + }, + { + "path": "audio-1499265084-headset.flac", + "question": "What's the first word that's alluding to 'place_name' in this sound recording?", + "choice_a": "downtown", + "choice_b": "there", + "choice_c": "new york", + "choice_d": "china", + "answer_gt": "downtown", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7035 + }, + { + "path": "audio-1498578754.flac", + "question": "In this audio, what's the first word that's reminiscent of 'event_name'?", + "choice_a": "to", + "choice_b": "pick up mark", + "choice_c": "soccer match", + "choice_d": "call mom", + "answer_gt": "pick up mark", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7036 + }, + { + "path": "audio-1501408019.flac", + "question": "What is the first word related to 'event_name' mentioned in this audio clip?", + "choice_a": "my", + "choice_b": "calendar", + "choice_c": "medical appointment", + "choice_d": "birthday", + "answer_gt": "medical appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7037 + }, + { + "path": "audio-1498564763.flac", + "question": "tell me the first word that's in relation to 'joke_type' in this audio?", + "choice_a": "funny", + "choice_b": "tell", + "choice_c": "good funny", + "choice_d": "funniest", + "answer_gt": "funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7038 + }, + { + "path": "audio-1498563824-headset.flac", + "question": "What is the first word related to 'personal_info' mentioned in this audio clip?", + "choice_a": "person's", + "choice_b": "add", + "choice_c": "phone numbers", + "choice_d": "email", + "answer_gt": "email", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7039 + }, + { + "path": "audio-1501756456-headset.flac", + "question": "What's the first word that's in reference to 'business_type' in this sound clip?", + "choice_a": "theater", + "choice_b": "tonight", + "choice_c": "cinema", + "choice_d": "movie", + "answer_gt": "movie", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7040 + }, + { + "path": "audio-1497622259-headset.flac", + "question": "Which word comes first in this audio that's tied to 'date'?", + "choice_a": "the twenty third of this month", + "choice_b": "third", + "choice_c": "fifth", + "choice_d": "friday", + "answer_gt": "the twenty third of this month", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7041 + }, + { + "path": "audio-1495378614.flac", + "question": "name the first word that's related to 'definition_word' in this audio?", + "choice_a": "of", + "choice_b": "velocity", + "choice_c": "smartphone circuit", + "choice_d": "forensic", + "answer_gt": "velocity", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7042 + }, + { + "path": "audio-1494417284.flac", + "question": "What's the first word that's associated with 'person' in this clip?", + "choice_a": "of", + "choice_b": "ajay devgn", + "choice_c": "kate's", + "choice_d": "clint eastwood", + "answer_gt": "ajay devgn", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7043 + }, + { + "path": "audio-1490800722.flac", + "question": "Which term that relates to 'place_name' is spoken first in this clip?", + "choice_a": "india's", + "choice_b": "tell", + "choice_c": "alaska", + "choice_d": "uk", + "answer_gt": "india's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7044 + }, + { + "path": "audio-1501752274-headset.flac", + "question": "detect the first mention of a word related to 'time' in this audio.", + "choice_a": "week's", + "choice_b": "next", + "choice_c": "fifteen minutes", + "choice_d": "next weeks", + "answer_gt": "next weeks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7045 + }, + { + "path": "audio-1434529054-headset.flac", + "question": "What's the first word that's alluding to 'relation' in this sound recording?", + "choice_a": "co worker", + "choice_b": "my", + "choice_c": "mom", + "choice_d": "coworker", + "answer_gt": "co worker", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7046 + }, + { + "path": "audio-1492783804-headset.flac", + "question": "Which word comes first in this audio that's tied to 'date'?", + "choice_a": "remind", + "choice_b": "tell", + "choice_c": "thirteenth june", + "choice_d": "this week", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7047 + }, + { + "path": "audio-1497861594-headset.flac", + "question": "Which word that's linked to 'transport_type' do you hear first in this recording?", + "choice_a": "half", + "choice_b": "leith", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7048 + }, + { + "path": "audio-1490261352.flac", + "question": "What's the first word that's reflective of 'person' in this sound clip?", + "choice_a": "john", + "choice_b": "contacts", + "choice_c": "john smith", + "choice_d": "margaret", + "answer_gt": "john smith", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7049 + }, + { + "path": "audio-1434542017-headset.flac", + "question": "What is the first word that's in connection with 'list_name' in this audio?", + "choice_a": "dollars", + "choice_b": "expenses", + "choice_c": "party time", + "choice_d": "kitchen", + "answer_gt": "expenses", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7050 + }, + { + "path": "audio-1489672370.flac", + "question": "What's the first word that's characteristic of 'place_name' in this clip?", + "choice_a": "the", + "choice_b": "new", + "choice_c": "new york", + "choice_d": "grand rapids mi", + "answer_gt": "new york", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7051 + }, + { + "path": "audio-1495372243-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'definition_word'?", + "choice_a": "a", + "choice_b": "hey", + "choice_c": "blanket", + "choice_d": "book", + "answer_gt": "book", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7052 + }, + { + "path": "audio-1500036584.flac", + "question": "What's the first word that's in reference to 'joke_type' in this sound clip?", + "choice_a": "me", + "choice_b": "tell", + "choice_c": "good", + "choice_d": "funny", + "answer_gt": "funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7053 + }, + { + "path": "audio-1495732777.flac", + "question": "Which word is introduced first in this audio that relates to 'radio_name'?", + "choice_a": "bbc", + "choice_b": "start", + "choice_c": "station that plays r. n. b.", + "choice_d": "nine hundred and ninety nine fm", + "answer_gt": "bbc", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7054 + }, + { + "path": "audio-1502101861-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'transport_type'?", + "choice_a": "when", + "choice_b": "train", + "choice_c": "yellow cab", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7055 + }, + { + "path": "audio-1499690442.flac", + "question": "What's the first term that pertains to 'transport_type' in this recording?", + "choice_a": "call", + "choice_b": "yellow", + "choice_c": "metro", + "choice_d": "yellow cab", + "answer_gt": "yellow cab", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7056 + }, + { + "path": "audio-1490801985.flac", + "question": "What's the premier word about 'weather_descriptor' in this sound clip?", + "choice_a": "take", + "choice_b": "should", + "choice_c": "jacket", + "choice_d": "raincoat", + "answer_gt": "raincoat", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7057 + }, + { + "path": "audio-1490263555.flac", + "question": "What's the first word that's in association with 'device_type' in this recording?", + "choice_a": "vacuum cleaner", + "choice_b": "vacuum", + "choice_c": "smart socket", + "choice_d": "robot roomba", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7058 + }, + { + "path": "audio-1501756067-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'definition_word'?", + "choice_a": "rock sand", + "choice_b": "rock", + "choice_c": "converse shoes", + "choice_d": "orange", + "answer_gt": "rock sand", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7059 + }, + { + "path": "audio-1502197118.flac", + "question": "In this audio, what's the first word you hear that's connected to 'event_name'?", + "choice_a": "on", + "choice_b": "appointment", + "choice_c": "dance class", + "choice_d": "local current events", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7060 + }, + { + "path": "audio-1495378364-headset.flac", + "question": "point out the first 'time'-related word uttered in this audio.", + "choice_a": "six am", + "choice_b": "new", + "choice_c": "seven am", + "choice_d": "midnight", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7061 + }, + { + "path": "audio-1498483483.flac", + "question": "What's the first word that's associated with 'media_type' in this clip?", + "choice_a": "mcdonald's", + "choice_b": "tweet", + "choice_c": "new york time's", + "choice_d": "audio book", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7062 + }, + { + "path": "audio-1490263640-headset.flac", + "question": "What's the first word that's representative of 'place_name' in this recording?", + "choice_a": "weather", + "choice_b": "right", + "choice_c": "orange tx", + "choice_d": "train station", + "answer_gt": "orange tx", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7063 + }, + { + "path": "audio-1501151326-headset.flac", + "question": "What's the first word that's connected to 'business_name' in this audio?", + "choice_a": "to", + "choice_b": "is", + "choice_c": "domino's", + "choice_d": "amazon", + "answer_gt": "domino's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7064 + }, + { + "path": "audio-1498578505.flac", + "question": "What's the first word that's indicative of 'order_type' in this audio?", + "choice_a": "provide", + "choice_b": "they", + "choice_c": "takeout", + "choice_d": "takeaway", + "answer_gt": "takeaway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7065 + }, + { + "path": "audio-1501773126.flac", + "question": "Which word that's linked to 'date' do you hear first in this recording?", + "choice_a": "the", + "choice_b": "music", + "choice_c": "yesterday", + "choice_d": "one thousand nine hundred and ninety", + "answer_gt": "one thousand nine hundred and ninety", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7066 + }, + { + "path": "audio-1488556185.flac", + "question": "Which word comes first in this audio that's tied to 'date'?", + "choice_a": "olly", + "choice_b": "i", + "choice_c": "tomorrow", + "choice_d": "this weekend", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7067 + }, + { + "path": "audio-1502896756-headset.flac", + "question": "Which word that's affiliated with 'player_setting' is spoken first in this clip?", + "choice_a": "play and shuffle", + "choice_b": "shuffle", + "choice_c": "put on", + "choice_d": "resume", + "answer_gt": "play and shuffle", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7068 + }, + { + "path": "audio-1502376537-headset.flac", + "question": "What is the first word in this clip that's relevant to 'email_folder'?", + "choice_a": "emails", + "choice_b": "have", + "choice_c": "my", + "choice_d": "inbox", + "answer_gt": "inbox", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7069 + }, + { + "path": "audio-1499089221.flac", + "question": "What is the opening word pertaining to 'business_name' mentioned here?", + "choice_a": "how", + "choice_b": "i. b. m.", + "choice_c": "@companyname", + "choice_d": "consumer service", + "answer_gt": "i. b. m.", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7070 + }, + { + "path": "audio-1497869438.flac", + "question": "detect the first mention of a word related to 'transport_name' in this audio.", + "choice_a": "ten thousand eight hundred and eighty seven", + "choice_b": "seven", + "choice_c": "rajdhani express", + "choice_d": "bagmati express", + "answer_gt": "ten thousand eight hundred and eighty seven", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7071 + }, + { + "path": "audio-1490359489-headset.flac", + "question": "What is the initial 'player_setting'-associated word you hear in this recording?", + "choice_a": "the", + "choice_b": "replay", + "choice_c": "move on", + "choice_d": "stop", + "answer_gt": "replay", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7072 + }, + { + "path": "audio-1496937888.flac", + "question": "What's the premier word about 'date' in this sound clip?", + "choice_a": "today", + "choice_b": "do", + "choice_c": "march eighteen", + "choice_d": "todays", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7073 + }, + { + "path": "audio-1501414606.flac", + "question": "What word is the first to reference 'transport_type' in this sound clip?", + "choice_a": "train", + "choice_b": "is", + "choice_c": "uberpool", + "choice_d": "metro", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7074 + }, + { + "path": "audio-1490018229.flac", + "question": "Which word linked to 'place_name' comes first in this recording?", + "choice_a": "are", + "choice_b": "chennai", + "choice_c": "us", + "choice_d": "new york city", + "answer_gt": "chennai", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7075 + }, + { + "path": "audio-1498566765.flac", + "question": "What's the first word that's suggestive of 'place_name' in this audio recording?", + "choice_a": "for", + "choice_b": "the", + "choice_c": "china", + "choice_d": "beijing", + "answer_gt": "china", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7076 + }, + { + "path": "audio-1501151535-headset.flac", + "question": "What's the first word that's suggestive of 'place_name' in this audio recording?", + "choice_a": "you", + "choice_b": "can", + "choice_c": "library", + "choice_d": "russia", + "answer_gt": "russia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7077 + }, + { + "path": "audio-1502200738-headset.flac", + "question": "Which word is the earliest mention of something related to 'house_place' in this audio?", + "choice_a": "shed", + "choice_b": "the", + "choice_c": "bathroom", + "choice_d": "office", + "answer_gt": "shed", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7078 + }, + { + "path": "audio-1502462223-headset.flac", + "question": "What's the first word that's reflective of 'radio_name' in this sound clip?", + "choice_a": "seven", + "choice_b": "just", + "choice_c": "station that plays r. n. b.", + "choice_d": "nine hundred and thirty seven fm", + "answer_gt": "nine hundred and thirty seven fm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7079 + }, + { + "path": "audio-1498566567.flac", + "question": "What word is the first to reference 'person' in this sound clip?", + "choice_a": "when", + "choice_b": "taylor", + "choice_c": "taylor lautner", + "choice_d": "ben's", + "answer_gt": "taylor lautner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7080 + }, + { + "path": "audio-1501756019.flac", + "question": "Which word comes first in this audio that's tied to 'ingredient'?", + "choice_a": "instead", + "choice_b": "can", + "choice_c": "oats", + "choice_d": "saffron", + "answer_gt": "saffron", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7081 + }, + { + "path": "audio-1500039533.flac", + "question": "Which word that's synonymous with 'relation' appears first in this audio?", + "choice_a": "boss", + "choice_b": "one", + "choice_c": "dad", + "choice_d": "mom's", + "answer_gt": "boss", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7082 + }, + { + "path": "audio-1498574145-headset.flac", + "question": "What's the first word that's associated with 'news_topic' in this clip?", + "choice_a": "the brexit", + "choice_b": "know", + "choice_c": "political", + "choice_d": "tornadoes", + "answer_gt": "the brexit", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7083 + }, + { + "path": "audio-1497884808.flac", + "question": "Which word that signifies 'date' is spoken first in this clip?", + "choice_a": "me", + "choice_b": "sunday", + "choice_c": "the nineteenth", + "choice_d": "everyday", + "answer_gt": "sunday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7084 + }, + { + "path": "audio-1497868983.flac", + "question": "Which word that's affiliated with 'transport_type' is spoken first in this clip?", + "choice_a": "uberpool", + "choice_b": "to", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "uberpool", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7085 + }, + { + "path": "audio-1434532388-headset.flac", + "question": "What is the first word in this clip that's relevant to 'music_genre'?", + "choice_a": "pop", + "choice_b": "fm", + "choice_c": "classic rock", + "choice_d": "techno", + "answer_gt": "pop", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7086 + }, + { + "path": "audio-1501691380.flac", + "question": "find the first word that's linked to 'time' in this sound clip?", + "choice_a": "there", + "choice_b": "ten am", + "choice_c": "seven am", + "choice_d": "five pm", + "answer_gt": "ten am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7087 + }, + { + "path": "audio-1494422748.flac", + "question": "What's the premier word about 'movie_name' in this sound clip?", + "choice_a": "kung fu panda three", + "choice_b": "panda", + "choice_c": "fu", + "choice_d": "fu", + "answer_gt": "kung fu panda three", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7088 + }, + { + "path": "audio-1502113445-headset.flac", + "question": "What is the first word that's associated with 'meal_type' in this audio?", + "choice_a": "at", + "choice_b": "every", + "choice_c": "dinner", + "choice_d": "lunch", + "answer_gt": "lunch", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7089 + }, + { + "path": "audio-1490356821-headset.flac", + "question": "What is the initial 'weather_descriptor'-associated word you hear in this recording?", + "choice_a": "sandals", + "choice_b": "wear", + "choice_c": "cold", + "choice_d": "hot", + "answer_gt": "sandals", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7090 + }, + { + "path": "audio-1502896651-headset.flac", + "question": "detect the first mention of a word related to 'place_name' in this audio.", + "choice_a": "how", + "choice_b": "large", + "choice_c": "library", + "choice_d": "alaska", + "answer_gt": "alaska", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7091 + }, + { + "path": "audio-1497029043.flac", + "question": "Which word comes first in this audio that's tied to 'event_name'?", + "choice_a": "meeting", + "choice_b": "nine", + "choice_c": "pre season baseball", + "choice_d": "order the turkey", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7092 + }, + { + "path": "audio-1502890414.flac", + "question": "What's the first word that's characteristic of 'ingredient' in this clip?", + "choice_a": "me", + "choice_b": "cook", + "choice_c": "oats", + "choice_d": "saffron", + "answer_gt": "oats", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7093 + }, + { + "path": "audio-1500035138-headset.flac", + "question": "Which word that's affiliated with 'event_name' is spoken first in this clip?", + "choice_a": "super", + "choice_b": "when", + "choice_c": "super bowl", + "choice_d": "medical appointment", + "answer_gt": "super bowl", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7094 + }, + { + "path": "audio-1494420545-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'food_type'?", + "choice_a": "how", + "choice_b": "cook", + "choice_c": "meatballs", + "choice_d": "red velvet cake", + "answer_gt": "meatballs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7095 + }, + { + "path": "audio-1502199282.flac", + "question": "What's the first word that comes up in this audio in relation to 'podcast_descriptor'?", + "choice_a": "next", + "choice_b": "play", + "choice_c": "rock songs having good rating", + "choice_d": "next podcast", + "answer_gt": "next podcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7096 + }, + { + "path": "audio-1490181170-headset.flac", + "question": "What is the opening word pertaining to 'game_name' mentioned here?", + "choice_a": "play", + "choice_b": "nfs", + "choice_c": "contra", + "choice_d": "clash of clans", + "answer_gt": "nfs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7097 + }, + { + "path": "audio-1494607589.flac", + "question": "What is the first word that's in connection with 'person' in this audio?", + "choice_a": "baby", + "choice_b": "hayathis", + "choice_c": "sarah", + "choice_d": "jeffrey burnette", + "answer_gt": "hayathis", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7098 + }, + { + "path": "audio-1490110359.flac", + "question": "What word relevant to 'event_name' is mentioned at the start of this audio?", + "choice_a": "pay days", + "choice_b": "add", + "choice_c": "church services", + "choice_d": "medicine", + "answer_gt": "pay days", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7099 + }, + { + "path": "audio-1501764802.flac", + "question": "What's the first term that pertains to 'media_type' in this recording?", + "choice_a": "twitter", + "choice_b": "a", + "choice_c": "facebook live", + "choice_d": "new york time's", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7100 + }, + { + "path": "audio-1490019370.flac", + "question": "What's the first word that's associated with 'date' in this clip?", + "choice_a": "york", + "choice_b": "please", + "choice_c": "everyday", + "choice_d": "todays", + "answer_gt": "todays", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7101 + }, + { + "path": "audio-1494417935.flac", + "question": "In this audio, what's the first word that's reminiscent of 'person'?", + "choice_a": "an", + "choice_b": "sally", + "choice_c": "kim kardashian", + "choice_d": "barack obama", + "answer_gt": "sally", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7102 + }, + { + "path": "audio--1506078152-headset.flac", + "question": "What's the first word that's tied to 'player_setting' mentioned in this audio?", + "choice_a": "place", + "choice_b": "resume", + "choice_c": "keep reading", + "choice_d": "skip", + "answer_gt": "resume", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7103 + }, + { + "path": "audio-1502895414-headset.flac", + "question": "What's the first word that's in reference to 'business_name' in this sound clip?", + "choice_a": "of", + "choice_b": "what", + "choice_c": "hdfc", + "choice_d": "walmart", + "answer_gt": "hdfc", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7104 + }, + { + "path": "audio-1501778149-headset.flac", + "question": "What is the first word that's associated with 'business_name' in this audio?", + "choice_a": "aircel", + "choice_b": "network", + "choice_c": "comcastcom", + "choice_d": "megatel", + "answer_gt": "aircel", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7105 + }, + { + "path": "audio-1489154013.flac", + "question": "Which word is the first to be related to 'food_type' in this recording?", + "choice_a": "the", + "choice_b": "pasta", + "choice_c": "pizza", + "choice_d": "pasta al dente", + "answer_gt": "pasta al dente", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7106 + }, + { + "path": "audio-1490110473-headset.flac", + "question": "Which word that signifies 'food_type' is spoken first in this clip?", + "choice_a": "pizza", + "choice_b": "from", + "choice_c": "pizza", + "choice_d": "pasta al dente", + "answer_gt": "pizza", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7107 + }, + { + "path": "audio--1504190166-headset.flac", + "question": "Which word that's linked to 'time' do you hear first in this recording?", + "choice_a": "fifteen minutes", + "choice_b": "in", + "choice_c": "a quarter to two", + "choice_d": "four", + "answer_gt": "fifteen minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7108 + }, + { + "path": "audio--1506079296-headset.flac", + "question": "What is the first word that's associated with 'person' in this audio?", + "choice_a": "barack", + "choice_b": "is", + "choice_c": "barack obama", + "choice_d": "kim kardashian", + "answer_gt": "barack obama", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7109 + }, + { + "path": "audio-1502891950.flac", + "question": "In this audio clip, what's the first word that pertains to 'player_setting'?", + "choice_a": "songs", + "choice_b": "shuffle", + "choice_c": "jump", + "choice_d": "hold", + "answer_gt": "shuffle", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7110 + }, + { + "path": "audio-1499266410.flac", + "question": "Which word that signifies 'food_type' is spoken first in this clip?", + "choice_a": "an", + "choice_b": "make", + "choice_c": "pasta", + "choice_d": "noodles", + "answer_gt": "noodles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7111 + }, + { + "path": "audio-1495734915.flac", + "question": "find the first word that's linked to 'person' in this sound clip?", + "choice_a": "george", + "choice_b": "clooney", + "choice_c": "mary's", + "choice_d": "george clooney", + "answer_gt": "george clooney", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7112 + }, + { + "path": "audio-1502891295.flac", + "question": "Which word that's linked to 'transport_type' do you hear first in this recording?", + "choice_a": "train", + "choice_b": "train", + "choice_c": "uberpool", + "choice_d": "yellow cab", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7113 + }, + { + "path": "audio-1492783602-headset.flac", + "question": "Which word is the earliest mention of something related to 'time' in this audio?", + "choice_a": "last", + "choice_b": "any", + "choice_c": "last hour", + "choice_d": "four", + "answer_gt": "last hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7114 + }, + { + "path": "audio-1498570806.flac", + "question": "What's the first word that's representative of 'artist_name' in this recording?", + "choice_a": "play", + "choice_b": "to", + "choice_c": "third day", + "choice_d": "sia's", + "answer_gt": "sia's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7115 + }, + { + "path": "audio-1488993353.flac", + "question": "Which word that signifies 'player_setting' is spoken first in this clip?", + "choice_a": "next", + "choice_b": "jump", + "choice_c": "playback", + "choice_d": "keep reading", + "answer_gt": "jump", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7116 + }, + { + "path": "audio-1499089971-headset.flac", + "question": "What's the first word that's indicative of 'podcast_name' in this audio?", + "choice_a": "from", + "choice_b": "green", + "choice_c": "sports podcast", + "choice_d": "order of the green hand", + "answer_gt": "order of the green hand", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7117 + }, + { + "path": "audio-1497886441-headset.flac", + "question": "What begins the list of 'date' words in this audio clip?", + "choice_a": "for", + "choice_b": "busy", + "choice_c": "rest of the year", + "choice_d": "tomorrow", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7118 + }, + { + "path": "audio-1501775422-headset.flac", + "question": "Which word that's synonymous with 'currency_name' appears first in this audio?", + "choice_a": "versus", + "choice_b": "the", + "choice_c": "u. s. d.", + "choice_d": "us dollar", + "answer_gt": "us dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7119 + }, + { + "path": "audio-1501776833-headset.flac", + "question": "What's the first word that's tied to 'place_name' mentioned in this audio?", + "choice_a": "a", + "choice_b": "to", + "choice_c": "australia", + "choice_d": "german", + "answer_gt": "german", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7120 + }, + { + "path": "audio-1501756057-headset.flac", + "question": "What's the first word that's representative of 'event_name' in this recording?", + "choice_a": "a", + "choice_b": "sports game", + "choice_c": "registration", + "choice_d": "church services", + "answer_gt": "sports game", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7121 + }, + { + "path": "audio-1497880250-headset.flac", + "question": "What begins the list of 'person' words in this audio clip?", + "choice_a": "any", + "choice_b": "new", + "choice_c": "elvis presley's", + "choice_d": "amy", + "answer_gt": "amy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7122 + }, + { + "path": "audio-1498571093-headset.flac", + "question": "What is the first word related to 'timeofday' mentioned in this audio clip?", + "choice_a": "are", + "choice_b": "there", + "choice_c": "morning", + "choice_d": "noon", + "answer_gt": "morning", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7123 + }, + { + "path": "audio-1498571298.flac", + "question": "What's the premier word about 'person' in this sound clip?", + "choice_a": "mahatma", + "choice_b": "where", + "choice_c": "alice", + "choice_d": "mahatma gandhi", + "answer_gt": "mahatma gandhi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7124 + }, + { + "path": "audio-1490106233.flac", + "question": "In this audio, what's the first word that's reminiscent of 'place_name'?", + "choice_a": "is", + "choice_b": "austin", + "choice_c": "miami", + "choice_d": "las vegas", + "answer_gt": "austin", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7125 + }, + { + "path": "audio-1501694857.flac", + "question": "Which word that's synonymous with 'event_name' appears first in this audio?", + "choice_a": "the", + "choice_b": "can", + "choice_c": "lee's birthday", + "choice_d": "meetings", + "answer_gt": "meetings", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7126 + }, + { + "path": "audio-1498574588.flac", + "question": "Which word that's synonymous with 'podcast_name' appears first in this audio?", + "choice_a": "want", + "choice_b": "i", + "choice_c": "sports podcast", + "choice_d": "disney", + "answer_gt": "sports podcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7127 + }, + { + "path": "audio-1501754637-headset.flac", + "question": "What's the first word that's tied to 'list_name' mentioned in this audio?", + "choice_a": "remove", + "choice_b": "stationary", + "choice_c": "old music", + "choice_d": "expenses", + "answer_gt": "stationary", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7128 + }, + { + "path": "audio-1492779900-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'device_type'?", + "choice_a": "robot", + "choice_b": "the", + "choice_c": "smart socket", + "choice_d": "robot roomba", + "answer_gt": "robot roomba", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7129 + }, + { + "path": "audio-1495370516.flac", + "question": "Which term that relates to 'media_type' is spoken first in this clip?", + "choice_a": "bbc", + "choice_b": "world", + "choice_c": "new york times", + "choice_d": "bbc world", + "answer_gt": "bbc world", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7130 + }, + { + "path": "audio-1492780234-headset.flac", + "question": "tell me the first word that's in relation to 'date' in this audio?", + "choice_a": "tomorrow", + "choice_b": "store", + "choice_c": "sunday", + "choice_d": "next week's", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7131 + }, + { + "path": "audio-1490006168-headset.flac", + "question": "tell me the first 'definition_word'-connected word in this audio.", + "choice_a": "about", + "choice_b": "all", + "choice_c": "hurricane", + "choice_d": "chair", + "answer_gt": "hurricane", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7132 + }, + { + "path": "audio-1499246103-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'media_type'?", + "choice_a": "tweet", + "choice_b": "product", + "choice_c": "bbc poll", + "choice_d": "insta", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7133 + }, + { + "path": "audio-1500038095.flac", + "question": "What is the initial 'weather_descriptor'-associated word you hear in this recording?", + "choice_a": "it", + "choice_b": "rain", + "choice_c": "sweater", + "choice_d": "driveway", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7134 + }, + { + "path": "audio-1497025325-headset.flac", + "question": "Which word that signifies 'event_name' is spoken first in this clip?", + "choice_a": "cleaning", + "choice_b": "eight", + "choice_c": "pick up my dry cleaning", + "choice_d": "supper", + "answer_gt": "pick up my dry cleaning", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7135 + }, + { + "path": "audio-1502200753.flac", + "question": "What word relevant to 'playlist_name' is mentioned at the start of this audio?", + "choice_a": "song", + "choice_b": "running", + "choice_c": "wacky", + "choice_d": "itunes library", + "answer_gt": "running", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7136 + }, + { + "path": "audio-1494607541-headset.flac", + "question": "Which term that relates to 'house_place' is spoken first in this clip?", + "choice_a": "in", + "choice_b": "green", + "choice_c": "living room", + "choice_d": "garage", + "answer_gt": "living room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7137 + }, + { + "path": "audio-1498571132.flac", + "question": "What's the first word that's associated with 'list_name' in this clip?", + "choice_a": "on", + "choice_b": "list", + "choice_c": "to dos", + "choice_d": "to do", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7138 + }, + { + "path": "audio-1501767885-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'radio_name'?", + "choice_a": "ninety", + "choice_b": "eight", + "choice_c": "i heart radio", + "choice_d": "eight hundred and ninety seven fm", + "answer_gt": "eight hundred and ninety seven fm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7139 + }, + { + "path": "audio-1490287389-headset.flac", + "question": "What is the first word in this clip that's relevant to 'weather_descriptor'?", + "choice_a": "open", + "choice_b": "open", + "choice_c": "temperature", + "choice_d": "sunscreen", + "answer_gt": "temperature", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7140 + }, + { + "path": "audio-1502308800-headset.flac", + "question": "What's the first word that's connected to 'person' in this audio?", + "choice_a": "eleven", + "choice_b": "calendar", + "choice_c": "ray", + "choice_d": "james and alice", + "answer_gt": "james and alice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7141 + }, + { + "path": "audio-1498572890.flac", + "question": "What's the first word that comes up in this audio in relation to 'news_topic'?", + "choice_a": "into", + "choice_b": "investigation into trump's ties", + "choice_c": "trump", + "choice_d": "environment", + "answer_gt": "investigation into trump's ties", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7142 + }, + { + "path": "audio-1501759249.flac", + "question": "What's the first word that's alluding to 'date' in this sound recording?", + "choice_a": "today", + "choice_b": "not", + "choice_c": "next week's", + "choice_d": "the twenty third of this month", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7143 + }, + { + "path": "audio-1490957397-headset.flac", + "question": "What's the first word that's reflective of 'person' in this sound clip?", + "choice_a": "bob", + "choice_b": "an", + "choice_c": "mariah carey", + "choice_d": "jessica's", + "answer_gt": "bob", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7144 + }, + { + "path": "audio-1490360188-headset.flac", + "question": "Which word is the earliest mention of something related to 'definition_word' in this audio?", + "choice_a": "is", + "choice_b": "definition", + "choice_c": "situation", + "choice_d": "forensic", + "answer_gt": "forensic", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7145 + }, + { + "path": "audio-1499695371.flac", + "question": "What's the first word mentioned in this audio that's about 'ingredient'?", + "choice_a": "what", + "choice_b": "used", + "choice_c": "saffron", + "choice_d": "oats", + "answer_gt": "saffron", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7146 + }, + { + "path": "audio-1500036874-headset.flac", + "question": "name the first word that's related to 'event_name' in this audio?", + "choice_a": "what", + "choice_b": "fringe", + "choice_c": "super bowl", + "choice_d": "dance class", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7147 + }, + { + "path": "audio-1490202252.flac", + "question": "Which word that signifies 'email_folder' is spoken first in this clip?", + "choice_a": "about", + "choice_b": "inbox", + "choice_c": "and", + "choice_d": "refresh", + "answer_gt": "inbox", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7148 + }, + { + "path": "audio--1506079613-headset.flac", + "question": "point out the first 'player_setting'-related word uttered in this audio.", + "choice_a": "keep", + "choice_b": "secret", + "choice_c": "skip", + "choice_d": "keep playing", + "answer_gt": "keep playing", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7149 + }, + { + "path": "audio-1498563767-headset.flac", + "question": "What's the first word that's associated with 'time' in this clip?", + "choice_a": "in", + "choice_b": "many", + "choice_c": "midnight", + "choice_d": "past three months", + "answer_gt": "midnight", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7150 + }, + { + "path": "audio-1499697062.flac", + "question": "tell me the first word that's in relation to 'house_place' in this audio?", + "choice_a": "hoover", + "choice_b": "the", + "choice_c": "office", + "choice_d": "hallway", + "answer_gt": "hallway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7151 + }, + { + "path": "audio-1494422741-headset.flac", + "question": "What's the earliest 'relation'-related term you hear in this clip?", + "choice_a": "is", + "choice_b": "from", + "choice_c": "friend", + "choice_d": "co worker", + "answer_gt": "friend", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7152 + }, + { + "path": "audio-1501608650.flac", + "question": "What's the first word that's associated with 'weather_descriptor' in this clip?", + "choice_a": "is", + "choice_b": "going", + "choice_c": "raining", + "choice_d": "rain", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7153 + }, + { + "path": "audio-1498485883-headset.flac", + "question": "What's the premier word about 'time' in this sound clip?", + "choice_a": "up", + "choice_b": "thirty minutes", + "choice_c": "last ten minutes", + "choice_d": "before", + "answer_gt": "thirty minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7154 + }, + { + "path": "audio-1498758847-headset.flac", + "question": "Which word is the earliest mention of something related to 'place_name' in this audio?", + "choice_a": "capital", + "choice_b": "kazakhstan", + "choice_c": "las vegas", + "choice_d": "international", + "answer_gt": "kazakhstan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7155 + }, + { + "path": "audio-1501150826-headset.flac", + "question": "find the first word that's linked to 'event_name' in this sound clip?", + "choice_a": "show", + "choice_b": "show", + "choice_c": "wedding", + "choice_d": "meetings", + "answer_gt": "meetings", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7156 + }, + { + "path": "audio-1495377894-headset.flac", + "question": "identify the initial word associated with 'place_name' in this audio.", + "choice_a": "got", + "choice_b": "us", + "choice_c": "the equator", + "choice_d": "fiji", + "answer_gt": "us", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7157 + }, + { + "path": "audio-1498564763-headset.flac", + "question": "In this sound clip, which word related to 'joke_type' is mentioned first?", + "choice_a": "funny", + "choice_b": "me", + "choice_c": "good", + "choice_d": "funniest", + "answer_gt": "funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7158 + }, + { + "path": "audio-1490704573.flac", + "question": "What's the first word that's characteristic of 'event_name' in this clip?", + "choice_a": "check the stew", + "choice_b": "the", + "choice_c": "parade", + "choice_d": "function", + "answer_gt": "check the stew", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7159 + }, + { + "path": "audio--1505404321.flac", + "question": "What is the first word that's associated with 'food_type' in this audio?", + "choice_a": "pizza", + "choice_b": "with", + "choice_c": "pasta", + "choice_d": "sushi", + "answer_gt": "pizza", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7160 + }, + { + "path": "audio-1501150397-headset.flac", + "question": "identify the initial word associated with 'food_type' in this audio.", + "choice_a": "risotto", + "choice_b": "know", + "choice_c": "stuffed peppers", + "choice_d": "sambar", + "answer_gt": "risotto", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7161 + }, + { + "path": "audio-1501152374.flac", + "question": "identify the initial word associated with 'list_name' in this audio.", + "choice_a": "add", + "choice_b": "grocery", + "choice_c": "stationary", + "choice_d": "activities", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7162 + }, + { + "path": "audio-1501686517.flac", + "question": "What's the first word that's characteristic of 'event_name' in this clip?", + "choice_a": "my", + "choice_b": "to", + "choice_c": "appointments", + "choice_d": "appointment", + "answer_gt": "appointments", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7163 + }, + { + "path": "audio--1504192655.flac", + "question": "What's the first word that comes up in this audio in relation to 'player_setting'?", + "choice_a": "resume", + "choice_b": "book", + "choice_c": "pause", + "choice_d": "continue reading", + "answer_gt": "resume", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7164 + }, + { + "path": "audio-1495369104-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'news_topic'?", + "choice_a": "presidential", + "choice_b": "the presidential election", + "choice_c": "trump", + "choice_d": "weather", + "answer_gt": "the presidential election", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7165 + }, + { + "path": "audio-1502099312.flac", + "question": "Which term that relates to 'media_type' is spoken first in this clip?", + "choice_a": "the", + "choice_b": "what", + "choice_c": "bbc poll", + "choice_d": "time of india", + "answer_gt": "bbc poll", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7166 + }, + { + "path": "audio-1501415524-headset.flac", + "question": "What is the initial 'audiobook_name'-associated word you hear in this recording?", + "choice_a": "and", + "choice_b": "play", + "choice_c": "harry potter", + "choice_d": "pride and prejudice", + "answer_gt": "pride and prejudice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7167 + }, + { + "path": "audio-1501415244.flac", + "question": "What's the earliest 'person'-related term you hear in this clip?", + "choice_a": "my", + "choice_b": "her", + "choice_c": "naty", + "choice_d": "kim kardashian", + "answer_gt": "naty", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7168 + }, + { + "path": "audio--1504192812.flac", + "question": "What's the earliest 'time'-related term you hear in this clip?", + "choice_a": "another hour", + "choice_b": "be", + "choice_c": "next weeks", + "choice_d": "three pm", + "answer_gt": "another hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7169 + }, + { + "path": "audio-1497622505.flac", + "question": "What's the first word that's reflective of 'event_name' in this sound clip?", + "choice_a": "on", + "choice_b": "lee's birthday", + "choice_c": "pre season baseball", + "choice_d": "wedding", + "answer_gt": "lee's birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7170 + }, + { + "path": "audio-1502300283.flac", + "question": "Which word is introduced first in this audio that relates to 'weather_descriptor'?", + "choice_a": "it", + "choice_b": "rain", + "choice_c": "warm socks", + "choice_d": "raincoat", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7171 + }, + { + "path": "audio-1498572707.flac", + "question": "What's the first word that's associated with 'relation' in this clip?", + "choice_a": "to", + "choice_b": "mother's", + "choice_c": "boss", + "choice_d": "grandma", + "answer_gt": "mother's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7172 + }, + { + "path": "audio-1499361225.flac", + "question": "spot the first word that hints at 'device_type' in this audio.", + "choice_a": "machine", + "choice_b": "you", + "choice_c": "coffee machine", + "choice_d": "okug", + "answer_gt": "coffee machine", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7173 + }, + { + "path": "audio-1501758762.flac", + "question": "In this audio, what's the first word that's reminiscent of 'place_name'?", + "choice_a": "the", + "choice_b": "in", + "choice_c": "beijing", + "choice_d": "sydney", + "answer_gt": "sydney", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7174 + }, + { + "path": "audio-1502893964.flac", + "question": "What's the first word that's tied to 'audiobook_name' mentioned in this audio?", + "choice_a": "listen", + "choice_b": "to", + "choice_c": "be more chill", + "choice_d": "twilight", + "answer_gt": "be more chill", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7175 + }, + { + "path": "audio-1499246231.flac", + "question": "What's the first word that's connected to 'place_name' in this audio?", + "choice_a": "kansas", + "choice_b": "check", + "choice_c": "paris", + "choice_d": "chennai", + "answer_gt": "kansas", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7176 + }, + { + "path": "audio-1498565994-headset.flac", + "question": "What's the first term that pertains to 'media_type' in this recording?", + "choice_a": "social", + "choice_b": "check", + "choice_c": "time of india", + "choice_d": "social networks", + "answer_gt": "social networks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7177 + }, + { + "path": "audio-1499267172-headset.flac", + "question": "What's the earliest 'food_type'-related term you hear in this clip?", + "choice_a": "know", + "choice_b": "for", + "choice_c": "pasta", + "choice_d": "meatballs", + "answer_gt": "pasta", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7178 + }, + { + "path": "audio-1502309660-headset.flac", + "question": "What's the earliest 'date'-related term you hear in this clip?", + "choice_a": "is", + "choice_b": "tuesday", + "choice_c": "next month", + "choice_d": "everyday", + "answer_gt": "tuesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7179 + }, + { + "path": "audio--1504195306.flac", + "question": "tell me the first word that's in relation to 'media_type' in this audio?", + "choice_a": "audiobook", + "choice_b": "open", + "choice_c": "audio book", + "choice_d": "facebook", + "answer_gt": "audiobook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7180 + }, + { + "path": "audio-1501767643-headset.flac", + "question": "Which term that relates to 'event_name' is spoken first in this clip?", + "choice_a": "party", + "choice_b": "please", + "choice_c": "medical appointment", + "choice_d": "supper", + "answer_gt": "party", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7181 + }, + { + "path": "audio-1434542317.flac", + "question": "What's the first word that's alluding to 'house_place' in this sound recording?", + "choice_a": "living room", + "choice_b": "dim", + "choice_c": "kitchen", + "choice_d": "my son's room", + "answer_gt": "living room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7182 + }, + { + "path": "audio-1488989256.flac", + "question": "Which word linked to 'coffee_type' comes first in this recording?", + "choice_a": "a", + "choice_b": "can", + "choice_c": "salted carmel flavored cream", + "choice_d": "creamy", + "answer_gt": "creamy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7183 + }, + { + "path": "audio-1499090087.flac", + "question": "In this audio clip, what's the first word that pertains to 'device_type'?", + "choice_a": "vacuum", + "choice_b": "start", + "choice_c": "vacuum cleaner", + "choice_d": "okug", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7184 + }, + { + "path": "audio-1501777644.flac", + "question": "What's the first word that's associated with 'player_setting' in this clip?", + "choice_a": "last", + "choice_b": "continue", + "choice_c": "shuffle", + "choice_d": "skip", + "answer_gt": "continue", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7185 + }, + { + "path": "audio-1490020976.flac", + "question": "What's the premier word about 'person' in this sound clip?", + "choice_a": "for", + "choice_b": "susan", + "choice_c": "chelsea", + "choice_d": "margaret", + "answer_gt": "susan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7186 + }, + { + "path": "audio-1492780117-headset.flac", + "question": "What's the first word that's reflective of 'date' in this sound clip?", + "choice_a": "a", + "choice_b": "will", + "choice_c": "one thousand nine hundred and ninety", + "choice_d": "today", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7187 + }, + { + "path": "audio-1502375164-headset.flac", + "question": "find the first word that's linked to 'event_name' in this sound clip?", + "choice_a": "bought", + "choice_b": "pre season baseball", + "choice_c": "medicine", + "choice_d": "sports game", + "answer_gt": "pre season baseball", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7188 + }, + { + "path": "audio-1488984246.flac", + "question": "What word relevant to 'event_name' is mentioned at the start of this audio?", + "choice_a": "match", + "choice_b": "be", + "choice_c": "medical appointment", + "choice_d": "soccer match", + "answer_gt": "soccer match", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7189 + }, + { + "path": "audio-1502377117-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'game_name'?", + "choice_a": "second", + "choice_b": "of", + "choice_c": "contra", + "choice_d": "poker", + "answer_gt": "contra", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7190 + }, + { + "path": "audio-1498566599-headset.flac", + "question": "What's the earliest 'date'-related term you hear in this clip?", + "choice_a": "off", + "choice_b": "rest of the year", + "choice_c": "Friday", + "choice_d": "friday", + "answer_gt": "rest of the year", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7191 + }, + { + "path": "audio-1502309334-headset.flac", + "question": "What's the first word in this audio that has a connection to 'date'?", + "choice_a": "me", + "choice_b": "today's", + "choice_c": "this weekend", + "choice_d": "todays", + "answer_gt": "today's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7192 + }, + { + "path": "audio-1497880821-headset.flac", + "question": "In this sound clip, which word related to 'date' is mentioned first?", + "choice_a": "new", + "choice_b": "give", + "choice_c": "todays", + "choice_d": "this year", + "answer_gt": "todays", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7193 + }, + { + "path": "audio-1501686813-headset.flac", + "question": "Which word linked to 'currency_name' comes first in this recording?", + "choice_a": "is", + "choice_b": "us", + "choice_c": "us dollar", + "choice_d": "u. s. d.", + "answer_gt": "us", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7194 + }, + { + "path": "audio-1497869077-headset.flac", + "question": "In this audio, what's the first word you hear that's connected to 'person'?", + "choice_a": "new", + "choice_b": "allearsnet", + "choice_c": "steph curry", + "choice_d": "mariah carey", + "answer_gt": "allearsnet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7195 + }, + { + "path": "audio-1497271972-headset.flac", + "question": "tell me the first word that's in relation to 'artist_name' in this audio?", + "choice_a": "lady gagas", + "choice_b": "what", + "choice_c": "justin bieber's", + "choice_d": "david bowie", + "answer_gt": "lady gagas", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7196 + }, + { + "path": "audio-1501691984.flac", + "question": "find the first word that's linked to 'device_type' in this sound clip?", + "choice_a": "the", + "choice_b": "vacuum", + "choice_c": "vacuum cleaner", + "choice_d": "wemo", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7197 + }, + { + "path": "audio-1490106767.flac", + "question": "Which word that's linked to 'music_genre' do you hear first in this recording?", + "choice_a": "rap", + "choice_b": "best", + "choice_c": "smooth jazz", + "choice_d": "rock", + "answer_gt": "rap", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7198 + }, + { + "path": "audio-1501756517-headset.flac", + "question": "What's the earliest 'coffee_type'-related term you hear in this clip?", + "choice_a": "please", + "choice_b": "make", + "choice_c": "creamy", + "choice_d": "without sweetener", + "answer_gt": "without sweetener", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7199 + }, + { + "path": "audio-1501766433.flac", + "question": "What's the first word that comes up in this audio in relation to 'event_name'?", + "choice_a": "shows", + "choice_b": "are", + "choice_c": "pre season baseball", + "choice_d": "gun shows", + "answer_gt": "gun shows", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7200 + }, + { + "path": "audio-1498565109.flac", + "question": "Which word is the first to be related to 'person' in this recording?", + "choice_a": "his", + "choice_b": "day", + "choice_c": "charlotte", + "choice_d": "obama", + "answer_gt": "obama", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7201 + }, + { + "path": "audio-1495371765-headset.flac", + "question": "Which word that's affiliated with 'artist_name' is spoken first in this clip?", + "choice_a": "celine dion", + "choice_b": "start", + "choice_c": "barenaked ladies", + "choice_d": "michael jackson", + "answer_gt": "celine dion", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7202 + }, + { + "path": "audio-1497449677.flac", + "question": "tell me the first 'person'-connected word in this audio.", + "choice_a": "old", + "choice_b": "mariah carey", + "choice_c": "pawel's", + "choice_d": "rita's", + "answer_gt": "mariah carey", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7203 + }, + { + "path": "audio-1501150404-headset.flac", + "question": "What's the first word in this audio that has a connection to 'weather_descriptor'?", + "choice_a": "stormy", + "choice_b": "if", + "choice_c": "raining", + "choice_d": "full cover jacket", + "answer_gt": "full cover jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7204 + }, + { + "path": "audio-1502890260-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'media_type'?", + "choice_a": "doctor", + "choice_b": "radio broadcast", + "choice_c": "cnn", + "choice_d": "new york time's", + "answer_gt": "radio broadcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7205 + }, + { + "path": "audio-1490355309.flac", + "question": "What's the first word that's characteristic of 'event_name' in this clip?", + "choice_a": "pick up my dry cleaning", + "choice_b": "remind", + "choice_c": "super bowl", + "choice_d": "dance class", + "answer_gt": "pick up my dry cleaning", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7206 + }, + { + "path": "audio-1502197983-headset.flac", + "question": "What's the first word that's representative of 'person' in this recording?", + "choice_a": "john's", + "choice_b": "play", + "choice_c": "lisa ann", + "choice_d": "kim kardashian", + "answer_gt": "john's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7207 + }, + { + "path": "audio-1498564460.flac", + "question": "In this audio, what's the first word that's reminiscent of 'date'?", + "choice_a": "eighteen", + "choice_b": "march eighteen", + "choice_c": "friday", + "choice_d": "yesterday", + "answer_gt": "march eighteen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7208 + }, + { + "path": "audio--1505405813-headset.flac", + "question": "tell me the first 'change_amount'-connected word in this audio.", + "choice_a": "fifty", + "choice_b": "volume", + "choice_c": "by fifty percent", + "choice_d": "twenty percent", + "answer_gt": "by fifty percent", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7209 + }, + { + "path": "audio-1490200110-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'place_name'?", + "choice_a": "new york city", + "choice_b": "to", + "choice_c": "russia", + "choice_d": "bazaar highway", + "answer_gt": "new york city", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7210 + }, + { + "path": "audio-1502200422.flac", + "question": "What's the first word that's representative of 'currency_name' in this recording?", + "choice_a": "what", + "choice_b": "price", + "choice_c": "us", + "choice_d": "dollar", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7211 + }, + { + "path": "audio-1501753618-headset.flac", + "question": "What's the first word that's suggestive of 'event_name' in this audio recording?", + "choice_a": "tal meeting", + "choice_b": "meeting", + "choice_c": "party", + "choice_d": "wash the windows", + "answer_gt": "tal meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7212 + }, + { + "path": "audio-1488622107.flac", + "question": "What is the opening word pertaining to 'date' mentioned here?", + "choice_a": "today's", + "choice_b": "what", + "choice_c": "yesterday", + "choice_d": "friday", + "answer_gt": "today's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7213 + }, + { + "path": "audio-1502216417.flac", + "question": "Which word that's evocative of 'place_name' is mentioned first in this audio?", + "choice_a": "about", + "choice_b": "india", + "choice_c": "jaipur", + "choice_d": "location", + "answer_gt": "india", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7214 + }, + { + "path": "audio-1501778586.flac", + "question": "In this audio, what's the first word you hear that's connected to 'event_name'?", + "choice_a": "you", + "choice_b": "for", + "choice_c": "meeting", + "choice_d": "wedding", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7215 + }, + { + "path": "audio-1490194892-headset.flac", + "question": "What is the first word that's in connection with 'event_name' in this audio?", + "choice_a": "haircut", + "choice_b": "clear", + "choice_c": "pick up mark", + "choice_d": "eric's bday", + "answer_gt": "haircut", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7216 + }, + { + "path": "audio-1498486088.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "the", + "choice_b": "alice", + "choice_c": "sarah", + "choice_d": "mahatma gandhi", + "answer_gt": "alice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7217 + }, + { + "path": "audio-1502099269.flac", + "question": "spot the first word that hints at 'place_name' in this audio.", + "choice_a": "traffic", + "choice_b": "commute", + "choice_c": "russia", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7218 + }, + { + "path": "audio-1498565442-headset.flac", + "question": "Which word is the earliest mention of something related to 'weather_descriptor' in this audio?", + "choice_a": "rain or snow", + "choice_b": "there", + "choice_c": "warm socks", + "choice_d": "jacket", + "answer_gt": "rain or snow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7219 + }, + { + "path": "audio-1498565483-headset.flac", + "question": "Which word that's linked to 'media_type' do you hear first in this recording?", + "choice_a": "who", + "choice_b": "is", + "choice_c": "new york time's", + "choice_d": "facebook live", + "answer_gt": "facebook live", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7220 + }, + { + "path": "audio-1490019948.flac", + "question": "tell me the first word that's in relation to 'house_place' in this audio?", + "choice_a": "son's", + "choice_b": "turn", + "choice_c": "bedroom", + "choice_d": "my son's room", + "answer_gt": "my son's room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7221 + }, + { + "path": "audio-1495376058-headset.flac", + "question": "name the first word that's related to 'event_name' in this audio?", + "choice_a": "meetings", + "choice_b": "tomorrow", + "choice_c": "medical appointment", + "choice_d": "easter", + "answer_gt": "meetings", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7222 + }, + { + "path": "audio-1502194796.flac", + "question": "What is the first word that's associated with 'media_type' in this audio?", + "choice_a": "from", + "choice_b": "anything", + "choice_c": "audio book", + "choice_d": "bob's news", + "answer_gt": "bob's news", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7223 + }, + { + "path": "audio-1502216574.flac", + "question": "What is the first word that's associated with 'date' in this audio?", + "choice_a": "need", + "choice_b": "week's", + "choice_c": "next week's", + "choice_d": "the nineteenth", + "answer_gt": "next week's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7224 + }, + { + "path": "audio-1490955473-headset.flac", + "question": "What's the first word that's in reference to 'transport_type' in this sound clip?", + "choice_a": "ticket", + "choice_b": "train", + "choice_c": "train", + "choice_d": "taxi", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7225 + }, + { + "path": "audio-1497868869.flac", + "question": "What word is the first to reference 'media_type' in this sound clip?", + "choice_a": "facebook", + "choice_b": "post", + "choice_c": "facebook", + "choice_d": "twitter", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7226 + }, + { + "path": "audio--1504190270-headset.flac", + "question": "tell me the first word that's in relation to 'player_setting' in this audio?", + "choice_a": "next", + "choice_b": "jump", + "choice_c": "pause", + "choice_d": "stop", + "answer_gt": "jump", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7227 + }, + { + "path": "audio-1499695891.flac", + "question": "In this audio, what's the first word that's reminiscent of 'transport_type'?", + "choice_a": "train", + "choice_b": "tell", + "choice_c": "cab", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7228 + }, + { + "path": "audio--1505406509-headset.flac", + "question": "In this audio, what's the first word you hear that's connected to 'change_amount'?", + "choice_a": "twenty", + "choice_b": "decrease", + "choice_c": "twenty percent", + "choice_d": "by fifty percent", + "answer_gt": "twenty percent", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7229 + }, + { + "path": "audio-1498574264-headset.flac", + "question": "What's the earliest 'place_name'-related term you hear in this clip?", + "choice_a": "india's", + "choice_b": "the", + "choice_c": "new york city", + "choice_d": "downtown", + "answer_gt": "india's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7230 + }, + { + "path": "audio-1489497417.flac", + "question": "tell me the first 'time'-connected word in this audio.", + "choice_a": "one hour", + "choice_b": "schedule", + "choice_c": "past three months", + "choice_d": "rest of the day", + "answer_gt": "one hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7231 + }, + { + "path": "audio-1499693864.flac", + "question": "What's the first word that's associated with 'event_name' in this clip?", + "choice_a": "appointment", + "choice_b": "an", + "choice_c": "doctor appointment", + "choice_d": "meeting", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7232 + }, + { + "path": "audio-1501687705-headset.flac", + "question": "detect the first mention of a word related to 'list_name' in this audio.", + "choice_a": "to", + "choice_b": "check", + "choice_c": "to do", + "choice_d": "old music", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7233 + }, + { + "path": "audio-1501693798.flac", + "question": "What's the first word that's in reference to 'music_genre' in this sound clip?", + "choice_a": "jazz", + "choice_b": "i", + "choice_c": "nineties rap", + "choice_d": "techno", + "answer_gt": "jazz", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7234 + }, + { + "path": "audio-1499694095-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'place_name'?", + "choice_a": "orange tx", + "choice_b": "having", + "choice_c": "home", + "choice_d": "jaipur", + "answer_gt": "orange tx", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7235 + }, + { + "path": "audio-1502895546.flac", + "question": "name the first word that's related to 'transport_name' in this audio?", + "choice_a": "ticket", + "choice_b": "a", + "choice_c": "bagmati express", + "choice_d": "rajdhani express", + "answer_gt": "rajdhani express", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7236 + }, + { + "path": "audio-1490368344-headset.flac", + "question": "What is the first word that's in connection with 'time' in this audio?", + "choice_a": "eight am", + "choice_b": "eight", + "choice_c": "before", + "choice_d": "another hour", + "answer_gt": "eight am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7237 + }, + { + "path": "audio-1495371114.flac", + "question": "What's the first word mentioned in this audio that's about 'person'?", + "choice_a": "how", + "choice_b": "k.", + "choice_c": "j. k. rowling", + "choice_d": "bob", + "answer_gt": "j. k. rowling", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7238 + }, + { + "path": "audio-1490705008.flac", + "question": "spot the first word that hints at 'device_type' in this audio.", + "choice_a": "smart", + "choice_b": "disable", + "choice_c": "robot vacuum cleaner", + "choice_d": "smart socket", + "answer_gt": "smart socket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7239 + }, + { + "path": "audio-1490354728.flac", + "question": "In this audio clip, what's the first word that pertains to 'currency_name'?", + "choice_a": "between", + "choice_b": "us", + "choice_c": "us dollar", + "choice_d": "pounds", + "answer_gt": "us", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7240 + }, + { + "path": "audio-1490966656.flac", + "question": "What is the first word in this clip that's relevant to 'player_setting'?", + "choice_a": "stop", + "choice_b": "stop play", + "choice_c": "start chapter five", + "choice_d": "keep reading", + "answer_gt": "stop play", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7241 + }, + { + "path": "audio-1492781098.flac", + "question": "What's the first word that's alluding to 'food_type' in this sound recording?", + "choice_a": "pizza", + "choice_b": "order", + "choice_c": "risotto", + "choice_d": "poor boy sandwich", + "answer_gt": "pizza", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7242 + }, + { + "path": "audio-1501754231.flac", + "question": "What's the first word that's in reference to 'color_type' in this sound clip?", + "choice_a": "in", + "choice_b": "red", + "choice_c": "blue", + "choice_d": "red", + "answer_gt": "red", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7243 + }, + { + "path": "audio-1498573243.flac", + "question": "What is the first word that's in connection with 'place_name' in this audio?", + "choice_a": "library", + "choice_b": "i", + "choice_c": "new york", + "choice_d": "sacramento", + "answer_gt": "library", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7244 + }, + { + "path": "audio-1501697393.flac", + "question": "Which word is the first to be related to 'date' in this recording?", + "choice_a": "to", + "choice_b": "event", + "choice_c": "tomorrow", + "choice_d": "wednesday", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7245 + }, + { + "path": "audio-1502891249-headset.flac", + "question": "What's the first word that's tied to 'media_type' mentioned in this audio?", + "choice_a": "latest", + "choice_b": "from", + "choice_c": "insta", + "choice_d": "xinhua", + "answer_gt": "xinhua", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7246 + }, + { + "path": "audio-1498573240-headset.flac", + "question": "Which word is the earliest mention of something related to 'player_setting' in this audio?", + "choice_a": "on", + "choice_b": "play", + "choice_c": "keep playing", + "choice_d": "shuffle", + "answer_gt": "shuffle", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7247 + }, + { + "path": "audio-1495369355-headset.flac", + "question": "What is the first word that's in connection with 'place_name' in this audio?", + "choice_a": "or", + "choice_b": "the", + "choice_c": "train station", + "choice_d": "city", + "answer_gt": "city", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7248 + }, + { + "path": "audio--1504190714.flac", + "question": "What is the first word that's associated with 'transport_type' in this audio?", + "choice_a": "book", + "choice_b": "ticket", + "choice_c": "metro", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7249 + }, + { + "path": "audio-1488985220.flac", + "question": "point out the first 'timeofday'-related word uttered in this audio.", + "choice_a": "noon", + "choice_b": "set", + "choice_c": "noon", + "choice_d": "this afternoon", + "answer_gt": "noon", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7250 + }, + { + "path": "audio-1488988251.flac", + "question": "In this audio clip, what's the first word that pertains to 'date'?", + "choice_a": "two", + "choice_b": "that", + "choice_c": "the twenty third of this month", + "choice_d": "two days", + "answer_gt": "two days", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7251 + }, + { + "path": "audio-1490357487-headset.flac", + "question": "spot the first word that hints at 'definition_word' in this audio.", + "choice_a": "the", + "choice_b": "motivation", + "choice_c": "velocity", + "choice_d": "television circuit", + "answer_gt": "motivation", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7252 + }, + { + "path": "audio-1497620604-headset.flac", + "question": "What is the initial 'meal_type'-associated word you hear in this recording?", + "choice_a": "for", + "choice_b": "dinner", + "choice_c": "lunch", + "choice_d": "dinner", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7253 + }, + { + "path": "audio-1490007473.flac", + "question": "What is the initial 'media_type'-associated word you hear in this recording?", + "choice_a": "me", + "choice_b": "get", + "choice_c": "facebook", + "choice_d": "audiobook", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7254 + }, + { + "path": "audio-1501775625.flac", + "question": "Which word that's synonymous with 'podcast_descriptor' appears first in this audio?", + "choice_a": "podcast", + "choice_b": "next podcast episode", + "choice_c": "rock songs having good rating", + "choice_d": "next episode", + "answer_gt": "next podcast episode", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7255 + }, + { + "path": "audio-1498484681.flac", + "question": "What's the first term that pertains to 'person' in this recording?", + "choice_a": "angeline", + "choice_b": "is", + "choice_c": "trump", + "choice_d": "jane's", + "answer_gt": "angeline", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7256 + }, + { + "path": "audio-1502100124.flac", + "question": "What's the first word that's indicative of 'event_name' in this audio?", + "choice_a": "lunch", + "choice_b": "on", + "choice_c": "lunch meeting", + "choice_d": "print the documents", + "answer_gt": "lunch meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7257 + }, + { + "path": "audio-1502100658-headset.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "when", + "choice_b": "tell", + "choice_c": "obama", + "choice_d": "hayathis", + "answer_gt": "obama", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7258 + }, + { + "path": "audio--1506521004.flac", + "question": "Which term that relates to 'time' is spoken first in this clip?", + "choice_a": "pm", + "choice_b": "five pm to six pm", + "choice_c": "twelve pm", + "choice_d": "rest of the day", + "answer_gt": "five pm to six pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7259 + }, + { + "path": "audio-1434533124-headset.flac", + "question": "What's the first word that's indicative of 'weather_descriptor' in this audio?", + "choice_a": "jacket", + "choice_b": "need", + "choice_c": "warm socks", + "choice_d": "driveway", + "answer_gt": "jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7260 + }, + { + "path": "audio-1498483329-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'person'?", + "choice_a": "kate's", + "choice_b": "write", + "choice_c": "julie smith", + "choice_d": "mahatma gandhi", + "answer_gt": "kate's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7261 + }, + { + "path": "audio-1499698214.flac", + "question": "What's the first word that's indicative of 'media_type' in this audio?", + "choice_a": "tweet", + "choice_b": "a", + "choice_c": "time of india", + "choice_d": "cnn", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7262 + }, + { + "path": "audio-1490960684-headset.flac", + "question": "What's the premier word about 'general_frequency' in this sound clip?", + "choice_a": "day", + "choice_b": "reminder", + "choice_c": "every day", + "choice_d": "hourly", + "answer_gt": "every day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7263 + }, + { + "path": "audio-1490703214-headset.flac", + "question": "What's the first word in this audio that has a connection to 'music_genre'?", + "choice_a": "on", + "choice_b": "to", + "choice_c": "techno", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7264 + }, + { + "path": "audio-1490106026-headset.flac", + "question": "What's the first word that's tied to 'ingredient' mentioned in this audio?", + "choice_a": "cook", + "choice_b": "me", + "choice_c": "saffron", + "choice_d": "oats", + "answer_gt": "oats", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7265 + }, + { + "path": "audio-1490104290-headset.flac", + "question": "What's the first word that's associated with 'weather_descriptor' in this clip?", + "choice_a": "to", + "choice_b": "jacket", + "choice_c": "sweater", + "choice_d": "raining", + "answer_gt": "jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7266 + }, + { + "path": "audio-1490370600.flac", + "question": "What's the first word that's indicative of 'list_name' in this audio?", + "choice_a": "add", + "choice_b": "grocery", + "choice_c": "activities", + "choice_d": "walmart list", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7267 + }, + { + "path": "audio-1490366492-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'event_name'?", + "choice_a": "for", + "choice_b": "what", + "choice_c": "supper", + "choice_d": "easter", + "answer_gt": "supper", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7268 + }, + { + "path": "audio--1504191116.flac", + "question": "Which word that's linked to 'time' do you hear first in this recording?", + "choice_a": "six am", + "choice_b": "alarms", + "choice_c": "rest of the day", + "choice_d": "past three months", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7269 + }, + { + "path": "audio-1492781292-headset.flac", + "question": "What's the first word that's alluding to 'email_address' in this sound recording?", + "choice_a": "to", + "choice_b": "newmail@gmail dot com", + "choice_c": "john@gmail dot com", + "choice_d": "juanpedro@gmail dot com", + "answer_gt": "newmail@gmail dot com", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7270 + }, + { + "path": "audio-1501772728-headset.flac", + "question": "What's the first term that pertains to 'place_name' in this recording?", + "choice_a": "the", + "choice_b": "minister", + "choice_c": "russia", + "choice_d": "downtown", + "answer_gt": "russia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7271 + }, + { + "path": "audio-1502113125.flac", + "question": "What's the first word that's representative of 'device_type' in this recording?", + "choice_a": "on", + "choice_b": "the", + "choice_c": "smart socket", + "choice_d": "robot vacuum cleaner", + "answer_gt": "robot vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7272 + }, + { + "path": "audio-1497191771-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'media_type'?", + "choice_a": "newest provider", + "choice_b": "you", + "choice_c": "bob's news", + "choice_d": "social networks", + "answer_gt": "newest provider", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7273 + }, + { + "path": "audio-1490359035.flac", + "question": "In this sound clip, which word related to 'playlist_name' is mentioned first?", + "choice_a": "my", + "choice_b": "summer", + "choice_c": "running", + "choice_d": "summer jams playlist", + "answer_gt": "summer jams playlist", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7274 + }, + { + "path": "audio-1498575804.flac", + "question": "find the first word that's linked to 'song_name' in this sound clip?", + "choice_a": "by", + "choice_b": "giants", + "choice_c": "poker face", + "choice_d": "bohemian raphsody", + "answer_gt": "giants", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7275 + }, + { + "path": "audio-1502197107.flac", + "question": "point out the first 'device_type'-related word uttered in this audio.", + "choice_a": "my", + "choice_b": "can", + "choice_c": "vacuum", + "choice_d": "coffee machine", + "answer_gt": "coffee machine", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7276 + }, + { + "path": "audio-1497883081-headset.flac", + "question": "In this audio, what's the first word you hear that's connected to 'time_zone'?", + "choice_a": "it", + "choice_b": "what", + "choice_c": "pacific", + "choice_d": "eastern", + "answer_gt": "pacific", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7277 + }, + { + "path": "audio-1497880264-headset.flac", + "question": "What's the first word that's alluding to 'device_type' in this sound recording?", + "choice_a": "run", + "choice_b": "the", + "choice_c": "vacuum", + "choice_d": "all the lights", + "answer_gt": "vacuum", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7278 + }, + { + "path": "audio-1502101541-headset.flac", + "question": "What's the first word that's in association with 'podcast_descriptor' in this recording?", + "choice_a": "play", + "choice_b": "episode", + "choice_c": "next podcast episode", + "choice_d": "next episode", + "answer_gt": "next episode", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7279 + }, + { + "path": "audio-1492779692-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'definition_word'?", + "choice_a": "you", + "choice_b": "describe", + "choice_c": "happy birthday", + "choice_d": "converse shoes", + "answer_gt": "happy birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7280 + }, + { + "path": "audio-1494421835-headset.flac", + "question": "Which word that's synonymous with 'player_setting' appears first in this audio?", + "choice_a": "the", + "choice_b": "on", + "choice_c": "put on", + "choice_d": "stop", + "answer_gt": "put on", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7281 + }, + { + "path": "audio-1434533127.flac", + "question": "Which word is the first to be related to 'device_type' in this recording?", + "choice_a": "please", + "choice_b": "pot", + "choice_c": "coffee pot", + "choice_d": "okug", + "answer_gt": "coffee pot", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7282 + }, + { + "path": "audio-1501778433.flac", + "question": "Which word that's evocative of 'artist_name' is mentioned first in this audio?", + "choice_a": "play", + "choice_b": "the", + "choice_c": "david bowie", + "choice_d": "the beatles", + "answer_gt": "the beatles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7283 + }, + { + "path": "audio-1490184465.flac", + "question": "Which term that relates to 'definition_word' is spoken first in this clip?", + "choice_a": "define", + "choice_b": "define", + "choice_c": "blanket", + "choice_d": "cylinder", + "answer_gt": "blanket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7284 + }, + { + "path": "audio-1501688946.flac", + "question": "Which word that's evocative of 'list_name' is mentioned first in this audio?", + "choice_a": "business", + "choice_b": "add", + "choice_c": "contact list", + "choice_d": "expenses", + "answer_gt": "contact list", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7285 + }, + { + "path": "audio-1501757749.flac", + "question": "What's the first word that's associated with 'relation' in this clip?", + "choice_a": "any", + "choice_b": "check", + "choice_c": "dad", + "choice_d": "boss", + "answer_gt": "dad", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7286 + }, + { + "path": "audio-1497269885-headset.flac", + "question": "What's the first word in this audio that has a connection to 'currency_name'?", + "choice_a": "and", + "choice_b": "euro", + "choice_c": "pounds", + "choice_d": "u. s. d.", + "answer_gt": "euro", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7287 + }, + { + "path": "audio-1497359901.flac", + "question": "What is the opening word pertaining to 'event_name' mentioned here?", + "choice_a": "fall", + "choice_b": "easter", + "choice_c": "dance class", + "choice_d": "church services", + "answer_gt": "easter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7288 + }, + { + "path": "audio--1504193140.flac", + "question": "What's the first word mentioned in this audio that's about 'player_setting'?", + "choice_a": "pause", + "choice_b": "the", + "choice_c": "replay", + "choice_d": "start chapter five", + "answer_gt": "pause", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7289 + }, + { + "path": "audio-1501415201.flac", + "question": "Which word is introduced first in this audio that relates to 'transport_type'?", + "choice_a": "first", + "choice_b": "metro", + "choice_c": "taxi", + "choice_d": "cab", + "answer_gt": "metro", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7290 + }, + { + "path": "audio--1506520652.flac", + "question": "tell me the first word that's in relation to 'order_type' in this audio?", + "choice_a": "they", + "choice_b": "can", + "choice_c": "delivery", + "choice_d": "takeaway", + "answer_gt": "takeaway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7291 + }, + { + "path": "audio-1501691814-headset.flac", + "question": "Which word that's linked to 'person' do you hear first in this recording?", + "choice_a": "morgan's", + "choice_b": "do", + "choice_c": "chelsea", + "choice_d": "george clooney", + "answer_gt": "morgan's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7292 + }, + { + "path": "audio-1434541552.flac", + "question": "In this sound clip, what's the first word that's relevant to 'person'?", + "choice_a": "john", + "choice_b": "me", + "choice_c": "john's", + "choice_d": "pink", + "answer_gt": "john", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7293 + }, + { + "path": "audio-1502101254-headset.flac", + "question": "Which word is the earliest mention of something related to 'event_name' in this audio?", + "choice_a": "i", + "choice_b": "pre season baseball", + "choice_c": "doctor appointment", + "choice_d": "meetings", + "answer_gt": "pre season baseball", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7294 + }, + { + "path": "audio-1502375435.flac", + "question": "Which word is the first to be related to 'transport_type' in this recording?", + "choice_a": "leaving", + "choice_b": "train", + "choice_c": "yellow cab", + "choice_d": "metro", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7295 + }, + { + "path": "audio-1499361320.flac", + "question": "Which word linked to 'device_type' comes first in this recording?", + "choice_a": "please", + "choice_b": "vacuum cleaner", + "choice_c": "smart plug", + "choice_d": "vacuum", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7296 + }, + { + "path": "audio-1502196155.flac", + "question": "What's the first word that's suggestive of 'event_name' in this audio recording?", + "choice_a": "do", + "choice_b": "about", + "choice_c": "wash the windows", + "choice_d": "fringe", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7297 + }, + { + "path": "audio-1501688180-headset.flac", + "question": "What's the first word that's tied to 'artist_name' mentioned in this audio?", + "choice_a": "barenaked", + "choice_b": "play", + "choice_c": "third day", + "choice_d": "barenaked ladies", + "answer_gt": "barenaked ladies", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7298 + }, + { + "path": "audio-1490195073-headset.flac", + "question": "What's the first word that's representative of 'playlist_name' in this recording?", + "choice_a": "first", + "choice_b": "itunes", + "choice_c": "summer jams playlist", + "choice_d": "itunes library", + "answer_gt": "itunes library", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7299 + }, + { + "path": "audio-1490284112-headset.flac", + "question": "What's the first word that's connected to 'date' in this audio?", + "choice_a": "tuesday", + "choice_b": "next", + "choice_c": "this week's", + "choice_d": "august fifteenth", + "answer_gt": "tuesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7300 + }, + { + "path": "audio-1497262995.flac", + "question": "Which word comes first in this audio that's tied to 'list_name'?", + "choice_a": "what", + "choice_b": "down", + "choice_c": "walmart list", + "choice_d": "old english", + "answer_gt": "walmart list", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7301 + }, + { + "path": "audio-1498483178.flac", + "question": "What's the first word that's alluding to 'place_name' in this sound recording?", + "choice_a": "tell", + "choice_b": "me", + "choice_c": "international", + "choice_d": "city", + "answer_gt": "international", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7302 + }, + { + "path": "audio-1501608879.flac", + "question": "What's the first word that's indicative of 'food_type' in this audio?", + "choice_a": "sambar", + "choice_b": "for", + "choice_c": "red velvet cake", + "choice_d": "risotto", + "answer_gt": "sambar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7303 + }, + { + "path": "audio-1497367199.flac", + "question": "What is the first word related to 'list_name' mentioned in this audio clip?", + "choice_a": "music", + "choice_b": "the", + "choice_c": "todo", + "choice_d": "old music", + "answer_gt": "old music", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7304 + }, + { + "path": "audio-1490198924-headset.flac", + "question": "What's the first word that's connected to 'radio_name' in this audio?", + "choice_a": "on", + "choice_b": "hundred", + "choice_c": "hot one hundred and five", + "choice_d": "nine hundred and ninety nine fm", + "answer_gt": "nine hundred and ninety nine fm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7305 + }, + { + "path": "audio-1492785058-headset.flac", + "question": "Which term that relates to 'place_name' is spoken first in this clip?", + "choice_a": "what", + "choice_b": "of", + "choice_c": "west virginia", + "choice_d": "delhi", + "answer_gt": "west virginia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7306 + }, + { + "path": "audio-1501758871.flac", + "question": "point out the first 'transport_type'-related word uttered in this audio.", + "choice_a": "train", + "choice_b": "a", + "choice_c": "metro", + "choice_d": "cab", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7307 + }, + { + "path": "audio-1488986587.flac", + "question": "What's the first word that's connected to 'house_place' in this audio?", + "choice_a": "the", + "choice_b": "turn", + "choice_c": "living room's", + "choice_d": "kitchen", + "answer_gt": "living room's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7308 + }, + { + "path": "audio--1505324882-headset.flac", + "question": "What begins the list of 'list_name' words in this audio clip?", + "choice_a": "groceries", + "choice_b": "delete", + "choice_c": "shopping", + "choice_d": "todo", + "answer_gt": "groceries", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7309 + }, + { + "path": "audio-1497025112.flac", + "question": "What is the first word that's associated with 'audiobook_name' in this audio?", + "choice_a": "read", + "choice_b": "olly", + "choice_c": "hunt for read october", + "choice_d": "pride and prejudice", + "answer_gt": "hunt for read october", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7310 + }, + { + "path": "audio-1497880047.flac", + "question": "In this audio, what's the first word that's reminiscent of 'weather_descriptor'?", + "choice_a": "chance", + "choice_b": "precipitation", + "choice_c": "full cover jacket", + "choice_d": "raining", + "answer_gt": "precipitation", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7311 + }, + { + "path": "audio-1502374829.flac", + "question": "What's the premier word about 'date' in this sound clip?", + "choice_a": "today", + "choice_b": "i", + "choice_c": "next week's", + "choice_d": "the nineteenth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7312 + }, + { + "path": "audio-1500036532.flac", + "question": "Which word comes first in this audio that's tied to 'date'?", + "choice_a": "this weekend", + "choice_b": "weekend", + "choice_c": "august fifteenth", + "choice_d": "two days", + "answer_gt": "this weekend", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7313 + }, + { + "path": "audio-1490106100-headset.flac", + "question": "What word relevant to 'date' is mentioned at the start of this audio?", + "choice_a": "show", + "choice_b": "this", + "choice_c": "this Friday", + "choice_d": "tomorrow", + "answer_gt": "this Friday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7314 + }, + { + "path": "audio-1501696059-headset.flac", + "question": "What's the first word in this audio that has a connection to 'person'?", + "choice_a": "to", + "choice_b": "is", + "choice_c": "jane", + "choice_d": "charlotte", + "answer_gt": "jane", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7315 + }, + { + "path": "audio-1502101658.flac", + "question": "What begins the list of 'weather_descriptor' words in this audio clip?", + "choice_a": "will", + "choice_b": "nice", + "choice_c": "sunscreen", + "choice_d": "jacket", + "answer_gt": "nice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7316 + }, + { + "path": "audio-1498571948.flac", + "question": "Which word that signifies 'media_type' is spoken first in this clip?", + "choice_a": "on", + "choice_b": "facebook", + "choice_c": "bob's news", + "choice_d": "audio book", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7317 + }, + { + "path": "audio-1502894512.flac", + "question": "What's the first word that comes up in this audio in relation to 'media_type'?", + "choice_a": "post", + "choice_b": "this", + "choice_c": "facebook", + "choice_d": "audio book", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7318 + }, + { + "path": "audio-1501753652-headset.flac", + "question": "name the first word that's related to 'joke_type' in this audio?", + "choice_a": "what's", + "choice_b": "a", + "choice_c": "good", + "choice_d": "good funny", + "answer_gt": "good", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7319 + }, + { + "path": "audio-1490369043.flac", + "question": "Which word comes first in this audio that's tied to 'meal_type'?", + "choice_a": "is", + "choice_b": "ready", + "choice_c": "dinner", + "choice_d": "lunch", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7320 + }, + { + "path": "audio-1500898558-headset.flac", + "question": "What's the first word that's reflective of 'event_name' in this sound clip?", + "choice_a": "me", + "choice_b": "painting exhibition", + "choice_c": "car payment", + "choice_d": "lunch meeting", + "answer_gt": "painting exhibition", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7321 + }, + { + "path": "audio-1501696551-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'time'?", + "choice_a": "on", + "choice_b": "at", + "choice_c": "nine am", + "choice_d": "five pm", + "answer_gt": "nine am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7322 + }, + { + "path": "audio-1502375482.flac", + "question": "What is the first word in this clip that's relevant to 'meal_type'?", + "choice_a": "dinner", + "choice_b": "a", + "choice_c": "dinner", + "choice_d": "lunch", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7323 + }, + { + "path": "audio-1502377515.flac", + "question": "identify the initial word associated with 'coffee_type' in this audio.", + "choice_a": "without sweetener", + "choice_b": "please", + "choice_c": "salted carmel flavored cream", + "choice_d": "creamy", + "answer_gt": "without sweetener", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7324 + }, + { + "path": "audio-1490957057.flac", + "question": "In this audio, what's the first word that's reminiscent of 'media_type'?", + "choice_a": "facebook", + "choice_b": "happy", + "choice_c": "tweet", + "choice_d": "social networks", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7325 + }, + { + "path": "audio-1434541421-headset.flac", + "question": "Which word that signifies 'transport_name' is spoken first in this clip?", + "choice_a": "of", + "choice_b": "the", + "choice_c": "bagmati express", + "choice_d": "ten thousand eight hundred and eighty seven", + "answer_gt": "bagmati express", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7326 + }, + { + "path": "audio-1500387478.flac", + "question": "Which word linked to 'coffee_type' comes first in this recording?", + "choice_a": "salted", + "choice_b": "a", + "choice_c": "salted carmel flavored cream", + "choice_d": "without sweetener", + "answer_gt": "salted carmel flavored cream", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7327 + }, + { + "path": "audio-1497439727-headset.flac", + "question": "What's the first term that pertains to 'date' in this recording?", + "choice_a": "i", + "choice_b": "received", + "choice_c": "today", + "choice_d": "one thousand nine hundred and ninety", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7328 + }, + { + "path": "audio-1500978449-headset.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "compose", + "choice_b": "ray", + "choice_c": "sally", + "choice_d": "hayathis", + "answer_gt": "ray", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7329 + }, + { + "path": "audio-1502299633.flac", + "question": "Which word is the first to be related to 'date' in this recording?", + "choice_a": "forecast", + "choice_b": "for", + "choice_c": "week", + "choice_d": "Thursday", + "answer_gt": "week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7330 + }, + { + "path": "audio-1502100164.flac", + "question": "What's the first word that's representative of 'date' in this recording?", + "choice_a": "weather", + "choice_b": "this", + "choice_c": "the twenty third of this month", + "choice_d": "this week", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7331 + }, + { + "path": "audio-1501755827.flac", + "question": "What is the initial 'event_name'-associated word you hear in this recording?", + "choice_a": "appointment", + "choice_b": "two", + "choice_c": "dentist appointment", + "choice_d": "car payment", + "answer_gt": "dentist appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7332 + }, + { + "path": "audio-1490104028.flac", + "question": "tell me the first word that's in relation to 'house_place' in this audio?", + "choice_a": "off", + "choice_b": "bathroom", + "choice_c": "hallway", + "choice_d": "kitchen", + "answer_gt": "bathroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7333 + }, + { + "path": "audio-1494548237.flac", + "question": "What is the first word related to 'game_name' mentioned in this audio clip?", + "choice_a": "potter", + "choice_b": "me", + "choice_c": "poker", + "choice_d": "harry potter and the chamber of secrets", + "answer_gt": "harry potter and the chamber of secrets", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7334 + }, + { + "path": "audio-1502114385-headset.flac", + "question": "What's the first word that's tied to 'person' mentioned in this audio?", + "choice_a": "margaret", + "choice_b": "to", + "choice_c": "jack", + "choice_d": "lisa ann", + "answer_gt": "margaret", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7335 + }, + { + "path": "audio-1494548467.flac", + "question": "In this audio, what's the first word you hear that's connected to 'weather_descriptor'?", + "choice_a": "will", + "choice_b": "it", + "choice_c": "raining", + "choice_d": "rainy", + "answer_gt": "rainy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7336 + }, + { + "path": "audio-1501764456-headset.flac", + "question": "What is the opening word pertaining to 'event_name' mentioned here?", + "choice_a": "winter", + "choice_b": "winter solstice", + "choice_c": "party", + "choice_d": "birthday", + "answer_gt": "winter solstice", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7337 + }, + { + "path": "audio-1499692951-headset.flac", + "question": "What is the first word related to 'joke_type' mentioned in this audio clip?", + "choice_a": "you", + "choice_b": "joke", + "choice_c": "baseball", + "choice_d": "funniest", + "answer_gt": "funniest", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7338 + }, + { + "path": "audio-1495371837-headset.flac", + "question": "name the first word that's related to 'music_descriptor' in this audio?", + "choice_a": "good", + "choice_b": "a", + "choice_c": "previous", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7339 + }, + { + "path": "audio-1488540656.flac", + "question": "Which word that's affiliated with 'food_type' is spoken first in this clip?", + "choice_a": "could", + "choice_b": "sushi", + "choice_c": "pav bhaji", + "choice_d": "poor boy sandwich", + "answer_gt": "sushi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7340 + }, + { + "path": "audio-1502375329-headset.flac", + "question": "find the first word that's linked to 'list_name' in this sound clip?", + "choice_a": "put", + "choice_b": "grocery", + "choice_c": "old english", + "choice_d": "grocery", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7341 + }, + { + "path": "audio--1504197832-headset.flac", + "question": "In this sound clip, which word related to 'audiobook_name' is mentioned first?", + "choice_a": "play", + "choice_b": "biography of jackie kennedy", + "choice_c": "twilight", + "choice_d": "pride and prejudice", + "answer_gt": "biography of jackie kennedy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7342 + }, + { + "path": "audio-1490202455-headset.flac", + "question": "tell me the first word that's in relation to 'artist_name' in this audio?", + "choice_a": "songs", + "choice_b": "major lazer", + "choice_c": "barenaked ladies", + "choice_d": "adele's", + "answer_gt": "major lazer", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7343 + }, + { + "path": "audio--1506079591.flac", + "question": "In this sound clip, which word related to 'time' is mentioned first?", + "choice_a": "ten", + "choice_b": "get", + "choice_c": "seven am", + "choice_d": "a quarter to two", + "answer_gt": "ten", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7344 + }, + { + "path": "audio--1504197574-headset.flac", + "question": "Which word comes first in this audio that's tied to 'player_setting'?", + "choice_a": "start", + "choice_b": "playback", + "choice_c": "replay", + "choice_d": "keep playing", + "answer_gt": "start", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7345 + }, + { + "path": "audio-1490202302.flac", + "question": "What's the first word that's alluding to 'event_name' in this sound recording?", + "choice_a": "order the turkey", + "choice_b": "turkey", + "choice_c": "soccer match", + "choice_d": "wedding", + "answer_gt": "order the turkey", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7346 + }, + { + "path": "audio-1497869722-headset.flac", + "question": "identify the initial word associated with 'date' in this audio.", + "choice_a": "is", + "choice_b": "today", + "choice_c": "weekdays", + "choice_d": "monday the twentieth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7347 + }, + { + "path": "audio-1490106265-headset.flac", + "question": "What's the first word that's representative of 'email_address' in this recording?", + "choice_a": "juanpedro@gmail", + "choice_b": "dot", + "choice_c": "newmail@gmail dot com", + "choice_d": "juanpedro@gmail dot com", + "answer_gt": "juanpedro@gmail dot com", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7348 + }, + { + "path": "audio-1500389159-headset.flac", + "question": "In this audio, what's the first word you hear that's connected to 'transport_type'?", + "choice_a": "taxi", + "choice_b": "leith", + "choice_c": "cab", + "choice_d": "yellow cab", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7349 + }, + { + "path": "audio-1501778826.flac", + "question": "What's the first term that pertains to 'weather_descriptor' in this recording?", + "choice_a": "i", + "choice_b": "raincoat", + "choice_c": "full cover jacket", + "choice_d": "cold", + "answer_gt": "raincoat", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7350 + }, + { + "path": "audio-1499266681-headset.flac", + "question": "point out the first 'person'-related word uttered in this audio.", + "choice_a": "who", + "choice_b": "cassy", + "choice_c": "steph curry", + "choice_d": "jessica snout's", + "answer_gt": "cassy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7351 + }, + { + "path": "audio-1495369380-headset.flac", + "question": "What's the first word that's associated with 'game_name' in this clip?", + "choice_a": "game", + "choice_b": "a", + "choice_c": "contra", + "choice_d": "nfs", + "answer_gt": "contra", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7352 + }, + { + "path": "audio--1504192922-headset.flac", + "question": "What's the earliest 'media_type'-related term you hear in this clip?", + "choice_a": "a", + "choice_b": "audio", + "choice_c": "newest provider", + "choice_d": "audio book", + "answer_gt": "audio book", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7353 + }, + { + "path": "audio-1498483119-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'food_type'?", + "choice_a": "to", + "choice_b": "i", + "choice_c": "pasta", + "choice_d": "stuffed peppers", + "answer_gt": "stuffed peppers", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7354 + }, + { + "path": "audio-1501772111.flac", + "question": "Which word that's affiliated with 'place_name' is spoken first in this clip?", + "choice_a": "how", + "choice_b": "is", + "choice_c": "new york city", + "choice_d": "alaska", + "answer_gt": "alaska", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7355 + }, + { + "path": "audio-1490704573-headset.flac", + "question": "spot the first word that hints at 'event_name' in this audio.", + "choice_a": "at", + "choice_b": "check the stew", + "choice_c": "fringe", + "choice_d": "wedding", + "answer_gt": "check the stew", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7356 + }, + { + "path": "audio-1494512684.flac", + "question": "What's the first word that's reflective of 'place_name' in this sound clip?", + "choice_a": "are", + "choice_b": "weather", + "choice_c": "world", + "choice_d": "orange tx", + "answer_gt": "orange tx", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7357 + }, + { + "path": "audio-1497450841.flac", + "question": "find the first word that's linked to 'place_name' in this sound clip?", + "choice_a": "la", + "choice_b": "any", + "choice_c": "beijing", + "choice_d": "canada", + "answer_gt": "la", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7358 + }, + { + "path": "audio-1498568186-headset.flac", + "question": "What begins the list of 'person' words in this audio clip?", + "choice_a": "did", + "choice_b": "charlotte", + "choice_c": "pawel's", + "choice_d": "joanna", + "answer_gt": "charlotte", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7359 + }, + { + "path": "audio-1497029566.flac", + "question": "What's the earliest 'news_topic'-related term you hear in this clip?", + "choice_a": "topic", + "choice_b": "new", + "choice_c": "politics", + "choice_d": "new york bridge proposal", + "answer_gt": "politics", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7360 + }, + { + "path": "audio--1505324814.flac", + "question": "Which word is the earliest mention of something related to 'time' in this audio?", + "choice_a": "in", + "choice_b": "sunrise", + "choice_c": "twelve pm", + "choice_d": "last hour", + "answer_gt": "sunrise", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7361 + }, + { + "path": "audio-1501752948.flac", + "question": "Which word that's synonymous with 'event_name' appears first in this audio?", + "choice_a": "is", + "choice_b": "two", + "choice_c": "doctor appointment", + "choice_d": "easter", + "answer_gt": "easter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7362 + }, + { + "path": "audio--1504191411.flac", + "question": "Which word that's synonymous with 'audiobook_name' appears first in this audio?", + "choice_a": "the", + "choice_b": "twilight", + "choice_c": "hunt for read october", + "choice_d": "tarzan", + "answer_gt": "twilight", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7363 + }, + { + "path": "audio-1498485816-headset.flac", + "question": "name the first word that's related to 'media_type' in this audio?", + "choice_a": "new york times", + "choice_b": "new", + "choice_c": "youtube", + "choice_d": "time of india", + "answer_gt": "new york times", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7364 + }, + { + "path": "audio-1490184465-headset.flac", + "question": "Which word comes first in this audio that's tied to 'definition_word'?", + "choice_a": "blanket", + "choice_b": "define", + "choice_c": "smartphone circuit", + "choice_d": "rock sand", + "answer_gt": "blanket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7365 + }, + { + "path": "audio-1497885530-headset.flac", + "question": "detect the first mention of a word related to 'relation' in this audio.", + "choice_a": "today", + "choice_b": "my", + "choice_c": "mother's", + "choice_d": "mom's", + "answer_gt": "mother's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7366 + }, + { + "path": "audio-1490201892-headset.flac", + "question": "Which word linked to 'event_name' comes first in this recording?", + "choice_a": "there", + "choice_b": "meetings", + "choice_c": "pre season baseball", + "choice_d": "medicine", + "answer_gt": "meetings", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7367 + }, + { + "path": "audio-1501772231.flac", + "question": "What's the first term that pertains to 'date' in this recording?", + "choice_a": "the", + "choice_b": "for", + "choice_c": "this week", + "choice_d": "a day before", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7368 + }, + { + "path": "audio-1501756232.flac", + "question": "What's the first word mentioned in this audio that's about 'person'?", + "choice_a": "send", + "choice_b": "reply", + "choice_c": "jack", + "choice_d": "amy", + "answer_gt": "jack", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7369 + }, + { + "path": "audio-1500038870.flac", + "question": "What's the first term that pertains to 'business_name' in this recording?", + "choice_a": "consumer service", + "choice_b": "report", + "choice_c": "pizza hut", + "choice_d": "terranova", + "answer_gt": "consumer service", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7370 + }, + { + "path": "audio-1497449617.flac", + "question": "What's the first word that's associated with 'news_topic' in this clip?", + "choice_a": "york", + "choice_b": "bridge", + "choice_c": "bitcoin price", + "choice_d": "new york bridge proposal", + "answer_gt": "new york bridge proposal", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7371 + }, + { + "path": "audio-1494608664-headset.flac", + "question": "What's the first word that's representative of 'date' in this recording?", + "choice_a": "next", + "choice_b": "weather", + "choice_c": "Thursday", + "choice_d": "wednesday", + "answer_gt": "wednesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7372 + }, + { + "path": "audio-1490290958.flac", + "question": "In this audio, what's the first word you hear that's connected to 'transport_name'?", + "choice_a": "is", + "choice_b": "of", + "choice_c": "bagmati express", + "choice_d": "rajdhani express", + "answer_gt": "bagmati express", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7373 + }, + { + "path": "audio-1494607217.flac", + "question": "What is the first word in this clip that's relevant to 'business_name'?", + "choice_a": "done", + "choice_b": "have", + "choice_c": "amazon", + "choice_d": "megatel", + "answer_gt": "megatel", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7374 + }, + { + "path": "audio-1501691897.flac", + "question": "Which word that's affiliated with 'media_type' is spoken first in this clip?", + "choice_a": "retweet", + "choice_b": "post", + "choice_c": "twitter", + "choice_d": "podcasts", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7375 + }, + { + "path": "audio-1490800722-headset.flac", + "question": "What's the first word that's suggestive of 'place_name' in this audio recording?", + "choice_a": "tell", + "choice_b": "india's", + "choice_c": "paris", + "choice_d": "my place", + "answer_gt": "india's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7376 + }, + { + "path": "audio-1490193091.flac", + "question": "What is the initial 'artist_name'-associated word you hear in this recording?", + "choice_a": "one", + "choice_b": "david bowie", + "choice_c": "third day", + "choice_d": "michael jackson", + "answer_gt": "david bowie", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7377 + }, + { + "path": "audio--1504194641.flac", + "question": "What is the first word that's associated with 'audiobook_name' in this audio?", + "choice_a": "play", + "choice_b": "book", + "choice_c": "tarzan", + "choice_d": "the notebook", + "answer_gt": "tarzan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7378 + }, + { + "path": "audio-1498576222.flac", + "question": "Which word that's evocative of 'place_name' is mentioned first in this audio?", + "choice_a": "what", + "choice_b": "london", + "choice_c": "paris", + "choice_d": "uk", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7379 + }, + { + "path": "audio-1492780289.flac", + "question": "Which word comes first in this audio that's tied to 'app_name'?", + "choice_a": "start", + "choice_b": "pandora", + "choice_c": "start", + "choice_d": "start", + "answer_gt": "pandora", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7380 + }, + { + "path": "audio-1501768348-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'event_name'?", + "choice_a": "know", + "choice_b": "fringe", + "choice_c": "wedding", + "choice_d": "gun shows", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7381 + }, + { + "path": "audio-1490956062-headset.flac", + "question": "What's the premier word about 'person' in this sound clip?", + "choice_a": "festival", + "choice_b": "mike", + "choice_c": "mary's", + "choice_d": "jerry", + "answer_gt": "mike", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7382 + }, + { + "path": "audio-1501696411.flac", + "question": "tell me the first 'date'-connected word in this audio.", + "choice_a": "is", + "choice_b": "searches", + "choice_c": "today", + "choice_d": "monday the twentieth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7383 + }, + { + "path": "audio-1501172973.flac", + "question": "What's the premier word about 'weather_descriptor' in this sound clip?", + "choice_a": "hot", + "choice_b": "how", + "choice_c": "rain or snow", + "choice_d": "sandals", + "answer_gt": "hot", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7384 + }, + { + "path": "audio-1502895872.flac", + "question": "What's the first word that's reflective of 'place_name' in this sound clip?", + "choice_a": "between", + "choice_b": "time", + "choice_c": "japan", + "choice_d": "one hundred and one", + "answer_gt": "japan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7385 + }, + { + "path": "audio-1502376424-headset.flac", + "question": "identify the initial word associated with 'food_type' in this audio.", + "choice_a": "long", + "choice_b": "turkey", + "choice_c": "tuna", + "choice_d": "pizza", + "answer_gt": "turkey", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7386 + }, + { + "path": "audio--1504197960.flac", + "question": "What's the first word in this audio that has a connection to 'time'?", + "choice_a": "six am", + "choice_b": "for", + "choice_c": "eight am", + "choice_d": "half an hour", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7387 + }, + { + "path": "audio-1502114299-headset.flac", + "question": "point out the first 'business_name'-related word uttered in this audio.", + "choice_a": "be", + "choice_b": "about", + "choice_c": "comcastcom", + "choice_d": "@companyname", + "answer_gt": "comcastcom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7388 + }, + { + "path": "audio-1497883127.flac", + "question": "What's the first word that comes up in this audio in relation to 'coffee_type'?", + "choice_a": "salted", + "choice_b": "of", + "choice_c": "salted carmel flavored cream", + "choice_d": "creamy", + "answer_gt": "salted carmel flavored cream", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7389 + }, + { + "path": "audio-1501765094-headset.flac", + "question": "What's the first word that's connected to 'podcast_descriptor' in this audio?", + "choice_a": "i", + "choice_b": "the", + "choice_c": "next podcast", + "choice_d": "rock songs having good rating", + "answer_gt": "next podcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7390 + }, + { + "path": "audio-1501687132.flac", + "question": "What is the opening word pertaining to 'music_genre' mentioned here?", + "choice_a": "my", + "choice_b": "play", + "choice_c": "reggae", + "choice_d": "nineties rap", + "answer_gt": "nineties rap", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7391 + }, + { + "path": "audio-1497884188.flac", + "question": "What is the initial 'person'-associated word you hear in this recording?", + "choice_a": "add", + "choice_b": "julie smith", + "choice_c": "morgan's", + "choice_d": "jane's", + "answer_gt": "julie smith", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7392 + }, + { + "path": "audio-1492785410-headset.flac", + "question": "What begins the list of 'person' words in this audio clip?", + "choice_a": "to", + "choice_b": "pawel's", + "choice_c": "billy crytals", + "choice_d": "sally", + "answer_gt": "pawel's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7393 + }, + { + "path": "audio-1502375004.flac", + "question": "What's the first term that pertains to 'food_type' in this recording?", + "choice_a": "one", + "choice_b": "noodles", + "choice_c": "turkey", + "choice_d": "pasta al dente", + "answer_gt": "noodles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7394 + }, + { + "path": "audio-1490261641.flac", + "question": "point out the first 'event_name'-related word uttered in this audio.", + "choice_a": "my", + "choice_b": "event", + "choice_c": "pre season baseball", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7395 + }, + { + "path": "audio--1506078275-headset.flac", + "question": "Which term that relates to 'game_name' is spoken first in this clip?", + "choice_a": "please", + "choice_b": "chess", + "choice_c": "battleship", + "choice_d": "contra", + "answer_gt": "chess", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7396 + }, + { + "path": "audio-1490291826.flac", + "question": "Which term that relates to 'currency_name' is spoken first in this clip?", + "choice_a": "day", + "choice_b": "for", + "choice_c": "u. s. d.", + "choice_d": "british pound", + "answer_gt": "u. s. d.", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7397 + }, + { + "path": "audio-1497884011.flac", + "question": "In this audio, what's the first word you hear that's connected to 'time'?", + "choice_a": "five am", + "choice_b": "five", + "choice_c": "three thirty pm", + "choice_d": "three pm", + "answer_gt": "five am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7398 + }, + { + "path": "audio-1490181170.flac", + "question": "Which word that signifies 'game_name' is spoken first in this clip?", + "choice_a": "play", + "choice_b": "high", + "choice_c": "poker", + "choice_d": "nfs", + "answer_gt": "nfs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7399 + }, + { + "path": "audio-1434542024.flac", + "question": "spot the first word that hints at 'business_name' in this audio.", + "choice_a": "i. b. m.", + "choice_b": "b.", + "choice_c": "mr pizza", + "choice_d": "income tax department", + "answer_gt": "i. b. m.", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7400 + }, + { + "path": "audio-1490287686-headset.flac", + "question": "What is the first word in this clip that's relevant to 'media_type'?", + "choice_a": "tweet", + "choice_b": "the", + "choice_c": "newest provider", + "choice_d": "twitter", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7401 + }, + { + "path": "audio-1489671862-headset.flac", + "question": "find the first word that's linked to 'person' in this sound clip?", + "choice_a": "do", + "choice_b": "phone", + "choice_c": "jessica snout's", + "choice_d": "kate's", + "answer_gt": "jessica snout's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7402 + }, + { + "path": "audio-1497883127-headset.flac", + "question": "Which word is the first to be related to 'coffee_type' in this recording?", + "choice_a": "salted", + "choice_b": "make", + "choice_c": "salted carmel flavored cream", + "choice_d": "without sweetener", + "answer_gt": "salted carmel flavored cream", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7403 + }, + { + "path": "audio-1501753618.flac", + "question": "tell me the first 'event_name'-connected word in this audio.", + "choice_a": "on", + "choice_b": "tal", + "choice_c": "tal meeting", + "choice_d": "conference call", + "answer_gt": "tal meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7404 + }, + { + "path": "audio-1501685829.flac", + "question": "In this audio clip, what's the first word that pertains to 'news_topic'?", + "choice_a": "and", + "choice_b": "me", + "choice_c": "trump", + "choice_d": "share market ups and downs", + "answer_gt": "share market ups and downs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7405 + }, + { + "path": "audio-1498485303-headset.flac", + "question": "What's the first word that's in reference to 'radio_name' in this sound clip?", + "choice_a": "start", + "choice_b": "my", + "choice_c": "fm", + "choice_d": "i heart radio", + "answer_gt": "i heart radio", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7406 + }, + { + "path": "audio-1501158292.flac", + "question": "What is the first word that's associated with 'podcast_descriptor' in this audio?", + "choice_a": "next", + "choice_b": "the", + "choice_c": "next podcast episode", + "choice_d": "next episode", + "answer_gt": "next episode", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7407 + }, + { + "path": "audio-1498564395-headset.flac", + "question": "Which word linked to 'date' comes first in this recording?", + "choice_a": "to", + "choice_b": "how", + "choice_c": "today", + "choice_d": "monday", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7408 + }, + { + "path": "audio-1490195692-headset.flac", + "question": "tell me the first 'event_name'-connected word in this audio.", + "choice_a": "on", + "choice_b": "wedding", + "choice_c": "fringe", + "choice_d": "pick up my parents", + "answer_gt": "wedding", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7409 + }, + { + "path": "audio-1488974480.flac", + "question": "What is the first word that's in connection with 'device_type' in this audio?", + "choice_a": "on", + "choice_b": "the", + "choice_c": "vacuum cleaner", + "choice_d": "all the lights", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7410 + }, + { + "path": "audio-1502896526-headset.flac", + "question": "What word is the first to reference 'transport_type' in this sound clip?", + "choice_a": "what", + "choice_b": "to", + "choice_c": "train", + "choice_d": "uberpool", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7411 + }, + { + "path": "audio-1495733140-headset.flac", + "question": "What's the first word that's alluding to 'news_topic' in this sound recording?", + "choice_a": "me", + "choice_b": "info", + "choice_c": "cars", + "choice_d": "trump", + "answer_gt": "cars", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7412 + }, + { + "path": "audio-1498566746-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'food_type'?", + "choice_a": "into", + "choice_b": "red", + "choice_c": "stuffed peppers", + "choice_d": "red velvet cake", + "answer_gt": "red velvet cake", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7413 + }, + { + "path": "audio-1490181890-headset.flac", + "question": "In this sound clip, which word related to 'email_folder' is mentioned first?", + "choice_a": "in", + "choice_b": "messages", + "choice_c": "inbox", + "choice_d": "if", + "answer_gt": "inbox", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7414 + }, + { + "path": "audio--1504198760-headset.flac", + "question": "tell me the first word that's in relation to 'time' in this audio?", + "choice_a": "at", + "choice_b": "ten", + "choice_c": "three pm", + "choice_d": "ten", + "answer_gt": "ten", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7415 + }, + { + "path": "audio-1502299614-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'transport_type'?", + "choice_a": "train", + "choice_b": "how", + "choice_c": "yellow cab", + "choice_d": "cab", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7416 + }, + { + "path": "audio-1502218389.flac", + "question": "Which word that's synonymous with 'event_name' appears first in this audio?", + "choice_a": "a", + "choice_b": "meeting", + "choice_c": "wedding", + "choice_d": "function", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7417 + }, + { + "path": "audio-1490706367.flac", + "question": "What is the opening word pertaining to 'relation' mentioned here?", + "choice_a": "an", + "choice_b": "coworker", + "choice_c": "grandma", + "choice_d": "wife", + "answer_gt": "coworker", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7418 + }, + { + "path": "audio--1504197960-headset.flac", + "question": "Which word that's synonymous with 'time' appears first in this audio?", + "choice_a": "six am", + "choice_b": "six", + "choice_c": "half an hour", + "choice_d": "eight am", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7419 + }, + { + "path": "audio-1501693586-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'player_setting'?", + "choice_a": "chapter", + "choice_b": "start chapter five", + "choice_c": "hold", + "choice_d": "playback", + "answer_gt": "start chapter five", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7420 + }, + { + "path": "audio-1494548013.flac", + "question": "What is the opening word pertaining to 'business_type' mentioned here?", + "choice_a": "store", + "choice_b": "a", + "choice_c": "furniture store", + "choice_d": "stores", + "answer_gt": "furniture store", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7421 + }, + { + "path": "audio-1501765460.flac", + "question": "What's the first word that's reflective of 'house_place' in this sound clip?", + "choice_a": "the", + "choice_b": "hoover", + "choice_c": "kitchen", + "choice_d": "hallway", + "answer_gt": "hallway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7422 + }, + { + "path": "audio-1495371114-headset.flac", + "question": "What's the first term that pertains to 'person' in this recording?", + "choice_a": "j. k. rowling", + "choice_b": "j.", + "choice_c": "bruce lee", + "choice_d": "john smith", + "answer_gt": "j. k. rowling", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7423 + }, + { + "path": "audio-1502200532.flac", + "question": "What's the first word that's in association with 'order_type' in this recording?", + "choice_a": "order", + "choice_b": "a", + "choice_c": "takeout", + "choice_d": "takeaway", + "answer_gt": "takeaway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7424 + }, + { + "path": "audio-1498571841.flac", + "question": "What word relevant to 'event_name' is mentioned at the start of this audio?", + "choice_a": "alert", + "choice_b": "registration", + "choice_c": "super bowl", + "choice_d": "order the turkey", + "answer_gt": "registration", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7425 + }, + { + "path": "audio-1498571132-headset.flac", + "question": "What is the first word related to 'list_name' mentioned in this audio clip?", + "choice_a": "to", + "choice_b": "is", + "choice_c": "to do", + "choice_d": "contact list", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7426 + }, + { + "path": "audio-1490369848.flac", + "question": "Which word linked to 'artist_name' comes first in this recording?", + "choice_a": "joe prsaise", + "choice_b": "song", + "choice_c": "justin bieber's", + "choice_d": "third day", + "answer_gt": "joe prsaise", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7427 + }, + { + "path": "audio-1488549851.flac", + "question": "What's the first word that's in reference to 'playlist_name' in this sound clip?", + "choice_a": "rock list", + "choice_b": "from", + "choice_c": "wacky", + "choice_d": "itunes library", + "answer_gt": "rock list", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7428 + }, + { + "path": "audio--1505405690-headset.flac", + "question": "Which word that's linked to 'event_name' do you hear first in this recording?", + "choice_a": "alarm", + "choice_b": "dance", + "choice_c": "dance class", + "choice_d": "birthday", + "answer_gt": "dance class", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7429 + }, + { + "path": "audio-1502895749-headset.flac", + "question": "What's the first word that's associated with 'time' in this clip?", + "choice_a": "six thirty am", + "choice_b": "for", + "choice_c": "another hour", + "choice_d": "midnight", + "answer_gt": "six thirty am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7430 + }, + { + "path": "audio-1490008661-headset.flac", + "question": "point out the first 'time'-related word uttered in this audio.", + "choice_a": "me", + "choice_b": "before", + "choice_c": "fifteen minutes", + "choice_d": "another hour", + "answer_gt": "before", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7431 + }, + { + "path": "audio-1499698419.flac", + "question": "What is the first word that's in connection with 'date' in this audio?", + "choice_a": "have", + "choice_b": "i", + "choice_c": "today", + "choice_d": "the nineteenth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7432 + }, + { + "path": "audio--1504190951-headset.flac", + "question": "Which word comes first in this audio that's tied to 'audiobook_name'?", + "choice_a": "davinci", + "choice_b": "davinci code", + "choice_c": "hunt for read october", + "choice_d": "harry potter", + "answer_gt": "davinci code", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7433 + }, + { + "path": "audio-1488639471.flac", + "question": "Which word that's synonymous with 'event_name' appears first in this audio?", + "choice_a": "nine", + "choice_b": "at", + "choice_c": "fringe", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7434 + }, + { + "path": "audio-1434533005.flac", + "question": "What is the first word that's associated with 'place_name' in this audio?", + "choice_a": "canada", + "choice_b": "to", + "choice_c": "disney land", + "choice_d": "mexico", + "answer_gt": "canada", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7435 + }, + { + "path": "audio-1489497766-headset.flac", + "question": "What's the first word that's representative of 'place_name' in this recording?", + "choice_a": "this city", + "choice_b": "time", + "choice_c": "bay area", + "choice_d": "japan", + "answer_gt": "this city", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7436 + }, + { + "path": "audio-1502302456-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'time'?", + "choice_a": "six am", + "choice_b": "six", + "choice_c": "nine am", + "choice_d": "before", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7437 + }, + { + "path": "audio--1504197574.flac", + "question": "What's the first word that's indicative of 'player_setting' in this audio?", + "choice_a": "start", + "choice_b": "playback", + "choice_c": "stop play", + "choice_d": "replay", + "answer_gt": "start", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7438 + }, + { + "path": "audio-1501778342.flac", + "question": "What's the first word mentioned in this audio that's about 'personal_info'?", + "choice_a": "numbers", + "choice_b": "i", + "choice_c": "email", + "choice_d": "phone numbers", + "answer_gt": "phone numbers", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7439 + }, + { + "path": "audio-1501773215.flac", + "question": "What's the first word that's representative of 'weather_descriptor' in this recording?", + "choice_a": "me", + "choice_b": "should", + "choice_c": "raincoat", + "choice_d": "jacket", + "answer_gt": "raincoat", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7440 + }, + { + "path": "audio-1501414573.flac", + "question": "In this sound clip, what's the first word that's relevant to 'music_genre'?", + "choice_a": "play", + "choice_b": "classic rock", + "choice_c": "sad songs", + "choice_d": "rap", + "answer_gt": "classic rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7441 + }, + { + "path": "audio--1505404136-headset.flac", + "question": "Which word that signifies 'date' is spoken first in this clip?", + "choice_a": "today", + "choice_b": "calendar", + "choice_c": "fifth", + "choice_d": "two days", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7442 + }, + { + "path": "audio--1504192520-headset.flac", + "question": "identify the initial word associated with 'change_amount' in this audio.", + "choice_a": "twenty", + "choice_b": "decrease", + "choice_c": "by fifty percent", + "choice_d": "twenty percent", + "answer_gt": "twenty percent", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7443 + }, + { + "path": "audio-1494608628-headset.flac", + "question": "Which word is the earliest mention of something related to 'artist_name' in this audio?", + "choice_a": "barenaked ladies", + "choice_b": "barenaked", + "choice_c": "sia's", + "choice_d": "ella fitzgerals", + "answer_gt": "barenaked ladies", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7444 + }, + { + "path": "audio-1501753322.flac", + "question": "identify the initial word associated with 'media_type' in this audio.", + "choice_a": "facebook", + "choice_b": "set", + "choice_c": "bbc poll", + "choice_d": "bob's news", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7445 + }, + { + "path": "audio-1498758020-headset.flac", + "question": "What's the first word in this audio that has a connection to 'event_name'?", + "choice_a": "dentist appointment", + "choice_b": "will", + "choice_c": "car payment", + "choice_d": "painting exhibition", + "answer_gt": "dentist appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7446 + }, + { + "path": "audio-1488991837.flac", + "question": "Which word comes first in this audio that's tied to 'food_type'?", + "choice_a": "as", + "choice_b": "noodles", + "choice_c": "red velvet cake", + "choice_d": "stuffed peppers", + "answer_gt": "noodles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7447 + }, + { + "path": "audio-1502890260.flac", + "question": "Which term that relates to 'media_type' is spoken first in this clip?", + "choice_a": "radio broadcast", + "choice_b": "who", + "choice_c": "bob's news", + "choice_d": "audio book", + "answer_gt": "radio broadcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7448 + }, + { + "path": "audio-1490287389.flac", + "question": "What's the first word that's suggestive of 'weather_descriptor' in this audio recording?", + "choice_a": "open", + "choice_b": "temperature", + "choice_c": "precipitation", + "choice_d": "jacket", + "answer_gt": "temperature", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7449 + }, + { + "path": "audio-1501407222.flac", + "question": "Which word that's evocative of 'place_name' is mentioned first in this audio?", + "choice_a": "everest", + "choice_b": "tall", + "choice_c": "home", + "choice_d": "town", + "answer_gt": "everest", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7450 + }, + { + "path": "audio-1502195540.flac", + "question": "Which word is the first to be related to 'time' in this recording?", + "choice_a": "tomorrow", + "choice_b": "one hour", + "choice_c": "six am", + "choice_d": "last hour", + "answer_gt": "one hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7451 + }, + { + "path": "audio-1434541527-headset.flac", + "question": "find the first word that's linked to 'person' in this sound clip?", + "choice_a": "is", + "choice_b": "steph", + "choice_c": "steph curry", + "choice_d": "chelsea", + "answer_gt": "steph curry", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7452 + }, + { + "path": "audio--1505405316-headset.flac", + "question": "What is the first word related to 'date' mentioned in this audio clip?", + "choice_a": "of", + "choice_b": "alarm", + "choice_c": "tuesday", + "choice_d": "today", + "answer_gt": "tuesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7453 + }, + { + "path": "audio-1500037046.flac", + "question": "Which word is the first to be related to 'time' in this recording?", + "choice_a": "london", + "choice_b": "midnight", + "choice_c": "nine am", + "choice_d": "three thirty pm", + "answer_gt": "midnight", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7454 + }, + { + "path": "audio-1500040593.flac", + "question": "What's the first word that's in association with 'media_type' in this recording?", + "choice_a": "about", + "choice_b": "a", + "choice_c": "tweet", + "choice_d": "facebook account", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7455 + }, + { + "path": "audio-1497558578-headset.flac", + "question": "What is the first word related to 'date' mentioned in this audio clip?", + "choice_a": "next", + "choice_b": "next month", + "choice_c": "today", + "choice_d": "yesterday", + "answer_gt": "next month", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7456 + }, + { + "path": "audio-1495377720.flac", + "question": "tell me the first word that's in relation to 'date' in this audio?", + "choice_a": "today", + "choice_b": "my", + "choice_c": "next week's", + "choice_d": "tomorrows", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7457 + }, + { + "path": "audio--1504198714-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'time'?", + "choice_a": "seven am", + "choice_b": "seven", + "choice_c": "past three months", + "choice_d": "ten am", + "answer_gt": "seven am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7458 + }, + { + "path": "audio-1501151909.flac", + "question": "What is the opening word pertaining to 'date' mentioned here?", + "choice_a": "add", + "choice_b": "at", + "choice_c": "tomorrow", + "choice_d": "march eighteen", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7459 + }, + { + "path": "audio-1498645566.flac", + "question": "What is the first word in this clip that's relevant to 'date'?", + "choice_a": "date", + "choice_b": "the", + "choice_c": "tomorrows", + "choice_d": "Friday", + "answer_gt": "tomorrows", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7460 + }, + { + "path": "audio-1501158292-headset.flac", + "question": "What word is the first to reference 'podcast_descriptor' in this sound clip?", + "choice_a": "the", + "choice_b": "play", + "choice_c": "next episode", + "choice_d": "next podcast episode", + "answer_gt": "next episode", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7461 + }, + { + "path": "audio-1502891400-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'definition_word'?", + "choice_a": "please", + "choice_b": "of", + "choice_c": "caftan", + "choice_d": "cylinder", + "answer_gt": "cylinder", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7462 + }, + { + "path": "audio--1504194804.flac", + "question": "What's the first word that's representative of 'player_setting' in this recording?", + "choice_a": "start", + "choice_b": "chapter", + "choice_c": "start chapter five", + "choice_d": "keep playing", + "answer_gt": "start chapter five", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7463 + }, + { + "path": "audio-1490705079-headset.flac", + "question": "What is the initial 'date'-associated word you hear in this recording?", + "choice_a": "meetings", + "choice_b": "your", + "choice_c": "Friday", + "choice_d": "second march", + "answer_gt": "second march", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7464 + }, + { + "path": "audio-1498566376-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'email_address'?", + "choice_a": "john@gmail", + "choice_b": "dot", + "choice_c": "john@gmail dot com", + "choice_d": "macs@gmail dot com", + "answer_gt": "john@gmail dot com", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7465 + }, + { + "path": "audio-1490103354-headset.flac", + "question": "What's the first word that's connected to 'place_name' in this audio?", + "choice_a": "bay area", + "choice_b": "bay", + "choice_c": "mexico", + "choice_d": "california", + "answer_gt": "bay area", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7466 + }, + { + "path": "audio-1498574204.flac", + "question": "What word relevant to 'weather_descriptor' is mentioned at the start of this audio?", + "choice_a": "a", + "choice_b": "need", + "choice_c": "rain or snow", + "choice_d": "jacket", + "answer_gt": "jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7467 + }, + { + "path": "audio-1502216910-headset.flac", + "question": "What word is the first to reference 'audiobook_name' in this sound clip?", + "choice_a": "please", + "choice_b": "notebook", + "choice_c": "the notebook", + "choice_d": "twilight", + "answer_gt": "the notebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7468 + }, + { + "path": "audio-1501758519-headset.flac", + "question": "What word is the first to reference 'time' in this sound clip?", + "choice_a": "last", + "choice_b": "last hour", + "choice_c": "five am", + "choice_d": "six am", + "answer_gt": "last hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7469 + }, + { + "path": "audio-1490368631-headset.flac", + "question": "detect the first mention of a word related to 'game_name' in this audio.", + "choice_a": "olly", + "choice_b": "play", + "choice_c": "chess", + "choice_d": "poker", + "answer_gt": "poker", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7470 + }, + { + "path": "audio-1490006530.flac", + "question": "What's the first word in this audio that has a connection to 'event_name'?", + "choice_a": "meeting", + "choice_b": "derrick", + "choice_c": "birthday", + "choice_d": "soccer match", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7471 + }, + { + "path": "audio-1490706367-headset.flac", + "question": "find the first word that's linked to 'relation' in this sound clip?", + "choice_a": "to", + "choice_b": "start", + "choice_c": "coworker", + "choice_d": "mom's", + "answer_gt": "coworker", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7472 + }, + { + "path": "audio-1490705888.flac", + "question": "What word relevant to 'relation' is mentioned at the start of this audio?", + "choice_a": "my", + "choice_b": "be", + "choice_c": "co worker", + "choice_d": "boss", + "answer_gt": "boss", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7473 + }, + { + "path": "audio-1490367408-headset.flac", + "question": "What's the first term that pertains to 'person' in this recording?", + "choice_a": "eliot's", + "choice_b": "november", + "choice_c": "george eliot's", + "choice_d": "lisa ann", + "answer_gt": "george eliot's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7474 + }, + { + "path": "audio-1502375329.flac", + "question": "In this audio, what's the first word you hear that's connected to 'list_name'?", + "choice_a": "grocery", + "choice_b": "hamburger", + "choice_c": "todo", + "choice_d": "grocery", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7475 + }, + { + "path": "audio-1501158131.flac", + "question": "What's the first word that's associated with 'time' in this clip?", + "choice_a": "up", + "choice_b": "thirty minutes", + "choice_c": "one hour", + "choice_d": "six thirty am", + "answer_gt": "thirty minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7476 + }, + { + "path": "audio-1489154939.flac", + "question": "Which word that's affiliated with 'transport_type' is spoken first in this clip?", + "choice_a": "glasgow", + "choice_b": "to", + "choice_c": "train", + "choice_d": "metro", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7477 + }, + { + "path": "audio-1492785833-headset.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "clint eastwood", + "choice_b": "is", + "choice_c": "john's", + "choice_d": "john", + "answer_gt": "clint eastwood", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7478 + }, + { + "path": "audio-1497439650.flac", + "question": "What is the initial 'currency_name'-associated word you hear in this recording?", + "choice_a": "what", + "choice_b": "price", + "choice_c": "dolla", + "choice_d": "dollar", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7479 + }, + { + "path": "audio-1498573957.flac", + "question": "What's the first word that's reflective of 'person' in this sound clip?", + "choice_a": "from", + "choice_b": "the", + "choice_c": "james and alice", + "choice_d": "john doe", + "answer_gt": "john doe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7480 + }, + { + "path": "audio-1490182506.flac", + "question": "Which word is introduced first in this audio that relates to 'place_name'?", + "choice_a": "rate", + "choice_b": "exchange", + "choice_c": "china", + "choice_d": "bolshoi theatre", + "answer_gt": "china", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7481 + }, + { + "path": "audio-1490287686.flac", + "question": "Which term that relates to 'media_type' is spoken first in this clip?", + "choice_a": "what", + "choice_b": "at", + "choice_c": "tweet", + "choice_d": "youtube", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7482 + }, + { + "path": "audio-1501688010.flac", + "question": "What's the first word that's alluding to 'event_name' in this sound recording?", + "choice_a": "would", + "choice_b": "on", + "choice_c": "church services", + "choice_d": "birthday", + "answer_gt": "church services", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7483 + }, + { + "path": "audio-1501688128.flac", + "question": "tell me the first word that's in relation to 'date' in this audio?", + "choice_a": "march eighteen", + "choice_b": "for", + "choice_c": "week", + "choice_d": "tomorrow's", + "answer_gt": "march eighteen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7484 + }, + { + "path": "audio-1502375525-headset.flac", + "question": "What's the first word that's alluding to 'definition_word' in this sound recording?", + "choice_a": "orange", + "choice_b": "definitions", + "choice_c": "happy birthday", + "choice_d": "sphere", + "answer_gt": "orange", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7485 + }, + { + "path": "audio-1497438654-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'song_name'?", + "choice_a": "play", + "choice_b": "silent", + "choice_c": "thirteen", + "choice_d": "giants", + "answer_gt": "silent", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7486 + }, + { + "path": "audio--1504190218.flac", + "question": "detect the first mention of a word related to 'game_name' in this audio.", + "choice_a": "a", + "choice_b": "game", + "choice_c": "contra", + "choice_d": "parcheese game", + "answer_gt": "contra", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7487 + }, + { + "path": "audio-1490262550.flac", + "question": "What is the initial 'date'-associated word you hear in this recording?", + "choice_a": "my", + "choice_b": "today", + "choice_c": "friday", + "choice_d": "fifth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7488 + }, + { + "path": "audio-1497271972.flac", + "question": "Which word comes first in this audio that's tied to 'artist_name'?", + "choice_a": "popular", + "choice_b": "what", + "choice_c": "lady gagas", + "choice_d": "michael jackson", + "answer_gt": "lady gagas", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7489 + }, + { + "path": "audio-1498578015.flac", + "question": "tell me the first word that's in relation to 'device_type' in this audio?", + "choice_a": "coffee pot", + "choice_b": "my", + "choice_c": "wemo plug socket", + "choice_d": "lamps", + "answer_gt": "coffee pot", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7490 + }, + { + "path": "audio-1502374642.flac", + "question": "What's the first word that's tied to 'event_name' mentioned in this audio?", + "choice_a": "bought", + "choice_b": "pre season baseball", + "choice_c": "lunch meeting", + "choice_d": "fringe", + "answer_gt": "pre season baseball", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7491 + }, + { + "path": "audio-1490202025-headset.flac", + "question": "What is the first word that's associated with 'media_type' in this audio?", + "choice_a": "sensational", + "choice_b": "time of india", + "choice_c": "new york time's", + "choice_d": "twitter", + "answer_gt": "time of india", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7492 + }, + { + "path": "audio-1490261246.flac", + "question": "Which word is the first to be related to 'date' in this recording?", + "choice_a": "has", + "choice_b": "your", + "choice_c": "today", + "choice_d": "todays", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7493 + }, + { + "path": "audio-1488974379.flac", + "question": "What begins the list of 'color_type' words in this audio clip?", + "choice_a": "is", + "choice_b": "pink", + "choice_c": "blue", + "choice_d": "red ish", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7494 + }, + { + "path": "audio-1499265306.flac", + "question": "What's the first word that's tied to 'event_name' mentioned in this audio?", + "choice_a": "me", + "choice_b": "musical", + "choice_c": "gun shows", + "choice_d": "musical events", + "answer_gt": "musical events", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7495 + }, + { + "path": "audio-1497024566.flac", + "question": "Which word is the first to be related to 'time' in this recording?", + "choice_a": "in", + "choice_b": "been", + "choice_c": "last ten minutes", + "choice_d": "eight am", + "answer_gt": "last ten minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7496 + }, + { + "path": "audio-1499090348-headset.flac", + "question": "Which term that relates to 'definition_word' is spoken first in this clip?", + "choice_a": "chair", + "choice_b": "a", + "choice_c": "hurricane", + "choice_d": "sphere", + "answer_gt": "chair", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7497 + }, + { + "path": "audio-1498574258-headset.flac", + "question": "What's the first word that's representative of 'date' in this recording?", + "choice_a": "for", + "choice_b": "got", + "choice_c": "today", + "choice_d": "monday", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7498 + }, + { + "path": "audio-1501151511.flac", + "question": "What's the first word that's in association with 'place_name' in this recording?", + "choice_a": "london", + "choice_b": "tell", + "choice_c": "india's", + "choice_d": "chennai", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7499 + }, + { + "path": "audio-1497887117-headset.flac", + "question": "Which word is the earliest mention of something related to 'place_name' in this audio?", + "choice_a": "us", + "choice_b": "got", + "choice_c": "detroit", + "choice_d": "russia", + "answer_gt": "us", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7500 + }, + { + "path": "audio-1502197973.flac", + "question": "What's the first word mentioned in this audio that's about 'transport_type'?", + "choice_a": "to", + "choice_b": "cab", + "choice_c": "taxi", + "choice_d": "metro", + "answer_gt": "cab", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7501 + }, + { + "path": "audio-1500387041-headset.flac", + "question": "What's the first word that's connected to 'event_name' in this audio?", + "choice_a": "meeting", + "choice_b": "please", + "choice_c": "anniversary", + "choice_d": "tal meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7502 + }, + { + "path": "audio-1501773170.flac", + "question": "What's the first word in this audio that has a connection to 'device_type'?", + "choice_a": "turn", + "choice_b": "on", + "choice_c": "vacuum cleaner", + "choice_d": "robot roomba", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7503 + }, + { + "path": "audio-1492784009-headset.flac", + "question": "point out the first 'time'-related word uttered in this audio.", + "choice_a": "five pm", + "choice_b": "for", + "choice_c": "a quarter to two", + "choice_d": "midnight", + "answer_gt": "five pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7504 + }, + { + "path": "audio-1498572458-headset.flac", + "question": "What word relevant to 'person' is mentioned at the start of this audio?", + "choice_a": "elvis presley's", + "choice_b": "what", + "choice_c": "sally", + "choice_d": "pawel's", + "answer_gt": "elvis presley's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7505 + }, + { + "path": "audio-1490106425.flac", + "question": "Which word comes first in this audio that's tied to 'food_type'?", + "choice_a": "way", + "choice_b": "is", + "choice_c": "pasta al dente", + "choice_d": "risotto", + "answer_gt": "pasta al dente", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7506 + }, + { + "path": "audio--1505405792-headset.flac", + "question": "What's the first word that's representative of 'player_setting' in this recording?", + "choice_a": "the", + "choice_b": "to", + "choice_c": "keep reading", + "choice_d": "continue reading", + "answer_gt": "keep reading", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7507 + }, + { + "path": "audio-1501692862-headset.flac", + "question": "What's the first word in this audio that has a connection to 'person'?", + "choice_a": "i", + "choice_b": "details", + "choice_c": "bruce lee", + "choice_d": "george clooney", + "answer_gt": "bruce lee", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7508 + }, + { + "path": "audio-1499265927-headset.flac", + "question": "What is the opening word pertaining to 'meal_type' mentioned here?", + "choice_a": "at", + "choice_b": "today", + "choice_c": "lunch", + "choice_d": "dinner", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7509 + }, + { + "path": "audio-1497558439-headset.flac", + "question": "tell me the first word that's in relation to 'business_type' in this audio?", + "choice_a": "tell", + "choice_b": "are", + "choice_c": "clothing stores", + "choice_d": "stores", + "answer_gt": "clothing stores", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7510 + }, + { + "path": "audio-1501692445.flac", + "question": "Which term that relates to 'joke_type' is spoken first in this clip?", + "choice_a": "good funny", + "choice_b": "would", + "choice_c": "funniest", + "choice_d": "funny", + "answer_gt": "good funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7511 + }, + { + "path": "audio-1495378090-headset.flac", + "question": "What's the first term that pertains to 'app_name' in this recording?", + "choice_a": "start", + "choice_b": "pandora", + "choice_c": "start", + "choice_d": "start", + "answer_gt": "pandora", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7512 + }, + { + "path": "audio-1499696274-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'person'?", + "choice_a": "donald trump", + "choice_b": "donald", + "choice_c": "bruce", + "choice_d": "rita's", + "answer_gt": "donald trump", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7513 + }, + { + "path": "audio-1501696084-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'definition_word'?", + "choice_a": "situation", + "choice_b": "does", + "choice_c": "converse shoes", + "choice_d": "blanket", + "answer_gt": "situation", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7514 + }, + { + "path": "audio--1504197881.flac", + "question": "Which word is introduced first in this audio that relates to 'date'?", + "choice_a": "tomorrow", + "choice_b": "to", + "choice_c": "thirteenth june", + "choice_d": "tomorrows", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7515 + }, + { + "path": "audio-1502112180.flac", + "question": "In this audio, what's the first word that's reminiscent of 'currency_name'?", + "choice_a": "equals", + "choice_b": "dolla", + "choice_c": "u. s. d.", + "choice_d": "us", + "answer_gt": "dolla", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7516 + }, + { + "path": "audio-1489154029.flac", + "question": "What's the first word in this audio that has a connection to 'event_name'?", + "choice_a": "for", + "choice_b": "add", + "choice_c": "dance class", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7517 + }, + { + "path": "audio-1502308513.flac", + "question": "detect the first mention of a word related to 'relation' in this audio.", + "choice_a": "birthday", + "choice_b": "mom", + "choice_c": "dad", + "choice_d": "wife", + "answer_gt": "mom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7518 + }, + { + "path": "audio-1501688874.flac", + "question": "What's the first word mentioned in this audio that's about 'artist_name'?", + "choice_a": "justin bieber's", + "choice_b": "next", + "choice_c": "lady gagas", + "choice_d": "celine dion", + "answer_gt": "justin bieber's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7519 + }, + { + "path": "audio-1488536771.flac", + "question": "Which word is the earliest mention of something related to 'person' in this audio?", + "choice_a": "joanna", + "choice_b": "saturday", + "choice_c": "angeline", + "choice_d": "naty", + "answer_gt": "joanna", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7520 + }, + { + "path": "audio-1502302042.flac", + "question": "detect the first mention of a word related to 'color_type' in this audio.", + "choice_a": "lights", + "choice_b": "to", + "choice_c": "red ish", + "choice_d": "blue", + "answer_gt": "blue", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7521 + }, + { + "path": "audio-1494606899.flac", + "question": "What word is the first to reference 'news_topic' in this sound clip?", + "choice_a": "business", + "choice_b": "me", + "choice_c": "president trump", + "choice_d": "the brexit", + "answer_gt": "business", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7522 + }, + { + "path": "audio--1504197955.flac", + "question": "Which word that's synonymous with 'date' appears first in this audio?", + "choice_a": "tomorrow", + "choice_b": "the", + "choice_c": "a day before", + "choice_d": "yesterday", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7523 + }, + { + "path": "audio-1490357632-headset.flac", + "question": "What's the premier word about 'event_name' in this sound clip?", + "choice_a": "my", + "choice_b": "parents", + "choice_c": "pick up my parents", + "choice_d": "fringe", + "answer_gt": "pick up my parents", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7524 + }, + { + "path": "audio-1502374812.flac", + "question": "What's the first word that's tied to 'date' mentioned in this audio?", + "choice_a": "want", + "choice_b": "i", + "choice_c": "next week's", + "choice_d": "everyday", + "answer_gt": "everyday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7525 + }, + { + "path": "audio-1497860774-headset.flac", + "question": "What is the first word that's in connection with 'definition_word' in this audio?", + "choice_a": "describe", + "choice_b": "what", + "choice_c": "sphere", + "choice_d": "smartphone circuit", + "answer_gt": "sphere", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7526 + }, + { + "path": "audio-1492757265.flac", + "question": "What's the first word that's in reference to 'date' in this sound clip?", + "choice_a": "my", + "choice_b": "like", + "choice_c": "Thursday", + "choice_d": "today", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7527 + }, + { + "path": "audio--1504194828.flac", + "question": "What's the first word that's in reference to 'event_name' in this sound clip?", + "choice_a": "in", + "choice_b": "to have tablets", + "choice_c": "theme parks", + "choice_d": "appointment", + "answer_gt": "to have tablets", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7528 + }, + { + "path": "audio-1490802270.flac", + "question": "Which word linked to 'place_name' comes first in this recording?", + "choice_a": "time", + "choice_b": "in", + "choice_c": "new york", + "choice_d": "australia", + "answer_gt": "new york", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7529 + }, + { + "path": "audio-1494607506-headset.flac", + "question": "Which word that's synonymous with 'place_name' appears first in this audio?", + "choice_a": "world", + "choice_b": "the", + "choice_c": "india's", + "choice_d": "disney land", + "answer_gt": "world", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7530 + }, + { + "path": "audio-1502199023-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'podcast_descriptor'?", + "choice_a": "play", + "choice_b": "next", + "choice_c": "next podcast", + "choice_d": "started last night", + "answer_gt": "next podcast", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7531 + }, + { + "path": "audio-1490263507-headset.flac", + "question": "Which word that's evocative of 'event_name' is mentioned first in this audio?", + "choice_a": "on", + "choice_b": "second of next mont lands", + "choice_c": "pay days", + "choice_d": "church services", + "answer_gt": "second of next mont lands", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7532 + }, + { + "path": "audio--1505406284-headset.flac", + "question": "tell me the first word that's in relation to 'time' in this audio?", + "choice_a": "seven", + "choice_b": "alarm", + "choice_c": "midnight", + "choice_d": "seven am", + "answer_gt": "seven am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7533 + }, + { + "path": "audio-1494548366-headset.flac", + "question": "What's the first word that's alluding to 'business_name' in this sound recording?", + "choice_a": "samsung", + "choice_b": "on", + "choice_c": "twitter", + "choice_d": "aircel", + "answer_gt": "samsung", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7534 + }, + { + "path": "audio-1502375638.flac", + "question": "point out the first 'movie_type'-related word uttered in this audio.", + "choice_a": "any", + "choice_b": "have", + "choice_c": "suggestion", + "choice_d": "action", + "answer_gt": "action", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7535 + }, + { + "path": "audio-1498568236-headset.flac", + "question": "What's the first word that's indicative of 'audiobook_name' in this audio?", + "choice_a": "code", + "choice_b": "davinci", + "choice_c": "the notebook", + "choice_d": "davinci code", + "answer_gt": "davinci code", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7536 + }, + { + "path": "audio-1501773487-headset.flac", + "question": "spot the first word that hints at 'date' in this audio.", + "choice_a": "tomorrow's", + "choice_b": "ten", + "choice_c": "monday", + "choice_d": "a day before", + "answer_gt": "tomorrow's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7537 + }, + { + "path": "audio-1434539865.flac", + "question": "In this sound clip, what's the first word that's relevant to 'date'?", + "choice_a": "which", + "choice_b": "today", + "choice_c": "everyday", + "choice_d": "two days", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7538 + }, + { + "path": "audio-1497886836-headset.flac", + "question": "Which word is the first to be related to 'weather_descriptor' in this recording?", + "choice_a": "it", + "choice_b": "is", + "choice_c": "cold", + "choice_d": "rainy", + "answer_gt": "cold", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7539 + }, + { + "path": "audio-1490705033.flac", + "question": "What's the first word that's associated with 'definition_word' in this clip?", + "choice_a": "define", + "choice_b": "define", + "choice_c": "situation", + "choice_d": "flabbergasted", + "answer_gt": "flabbergasted", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7540 + }, + { + "path": "audio-1501415356.flac", + "question": "Which word that's affiliated with 'date' is spoken first in this clip?", + "choice_a": "what", + "choice_b": "today", + "choice_c": "friday", + "choice_d": "weekdays", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7541 + }, + { + "path": "audio-1495370936-headset.flac", + "question": "What's the first word that's alluding to 'event_name' in this sound recording?", + "choice_a": "reminder", + "choice_b": "twenty", + "choice_c": "to have tablets", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7542 + }, + { + "path": "audio-1497881933-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'date'?", + "choice_a": "fifth", + "choice_b": "what", + "choice_c": "rest of the year", + "choice_d": "fifth march", + "answer_gt": "fifth", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7543 + }, + { + "path": "audio-1502896346-headset.flac", + "question": "What is the first word related to 'place_name' mentioned in this audio clip?", + "choice_a": "current", + "choice_b": "my", + "choice_c": "bazaar highway", + "choice_d": "canada", + "answer_gt": "canada", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7544 + }, + { + "path": "audio-1502199190.flac", + "question": "Which word is the first to be related to 'joke_type' in this recording?", + "choice_a": "tell", + "choice_b": "funny", + "choice_c": "baseball", + "choice_d": "good funny", + "answer_gt": "funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7545 + }, + { + "path": "audio--1504197671.flac", + "question": "What's the first word that's tied to 'date' mentioned in this audio?", + "choice_a": "tomorrow", + "choice_b": "wake", + "choice_c": "week", + "choice_d": "Friday", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7546 + }, + { + "path": "audio-1501686888.flac", + "question": "What is the first word that's associated with 'app_name' in this audio?", + "choice_a": "start", + "choice_b": "start", + "choice_c": "pandora", + "choice_d": "start", + "answer_gt": "pandora", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7547 + }, + { + "path": "audio-1490193161.flac", + "question": "What's the first word that's reflective of 'house_place' in this sound clip?", + "choice_a": "to", + "choice_b": "bedroom", + "choice_c": "bedroom", + "choice_d": "office", + "answer_gt": "bedroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7548 + }, + { + "path": "audio-1488974241.flac", + "question": "What's the first word that's characteristic of 'artist_name' in this clip?", + "choice_a": "ella", + "choice_b": "get", + "choice_c": "third day", + "choice_d": "ella fitzgerals", + "answer_gt": "ella fitzgerals", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7549 + }, + { + "path": "audio-1501689852-headset.flac", + "question": "What is the first word in this clip that's relevant to 'person'?", + "choice_a": "plays", + "choice_b": "patrick stewart", + "choice_c": "elvis", + "choice_d": "morgan freeman", + "answer_gt": "patrick stewart", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7550 + }, + { + "path": "audio-1489672066-headset.flac", + "question": "Which word linked to 'podcast_descriptor' comes first in this recording?", + "choice_a": "show", + "choice_b": "of", + "choice_c": "next podcast", + "choice_d": "rock songs having good rating", + "answer_gt": "rock songs having good rating", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7551 + }, + { + "path": "audio-1494512463.flac", + "question": "identify the initial word associated with 'music_descriptor' in this audio.", + "choice_a": "some", + "choice_b": "my", + "choice_c": "good song", + "choice_d": "christian", + "answer_gt": "good song", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7552 + }, + { + "path": "audio-1502100939.flac", + "question": "What begins the list of 'media_type' words in this audio clip?", + "choice_a": "twitter", + "choice_b": "the", + "choice_c": "insta", + "choice_d": "time of india", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7553 + }, + { + "path": "audio-1502300492-headset.flac", + "question": "What is the first word that's in connection with 'date' in this audio?", + "choice_a": "want", + "choice_b": "coffee", + "choice_c": "everyday", + "choice_d": "this week", + "answer_gt": "everyday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7554 + }, + { + "path": "audio-1490264552-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'joke_type'?", + "choice_a": "for", + "choice_b": "a", + "choice_c": "funny", + "choice_d": "good", + "answer_gt": "funny", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7555 + }, + { + "path": "audio-1501766310-headset.flac", + "question": "Which word that signifies 'time' is spoken first in this clip?", + "choice_a": "in", + "choice_b": "mark", + "choice_c": "four", + "choice_d": "twelve pm", + "answer_gt": "four", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7556 + }, + { + "path": "audio-1502099786-headset.flac", + "question": "Which word that's evocative of 'business_name' is mentioned first in this audio?", + "choice_a": "income tax department", + "choice_b": "income", + "choice_c": "comcastcom", + "choice_d": "twitter", + "answer_gt": "income tax department", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7557 + }, + { + "path": "audio-1499692407.flac", + "question": "What is the first word in this clip that's relevant to 'date'?", + "choice_a": "yesterday", + "choice_b": "mailed", + "choice_c": "twenty first", + "choice_d": "everyday", + "answer_gt": "yesterday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7558 + }, + { + "path": "audio-1497862912-headset.flac", + "question": "What word is the first to reference 'house_place' in this sound clip?", + "choice_a": "the", + "choice_b": "room", + "choice_c": "living room", + "choice_d": "office", + "answer_gt": "living room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7559 + }, + { + "path": "audio-1502098944.flac", + "question": "What's the first word in this audio that has a connection to 'event_name'?", + "choice_a": "have", + "choice_b": "i", + "choice_c": "pick up my parents", + "choice_d": "appointment", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7560 + }, + { + "path": "audio--1504191159-headset.flac", + "question": "What's the first word that's in reference to 'event_name' in this sound clip?", + "choice_a": "in", + "choice_b": "to have tablets", + "choice_c": "eric's bday", + "choice_d": "gun shows", + "answer_gt": "to have tablets", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7561 + }, + { + "path": "audio-1497439251-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'time'?", + "choice_a": "me", + "choice_b": "half an hour", + "choice_c": "sunrise", + "choice_d": "five am", + "answer_gt": "half an hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7562 + }, + { + "path": "audio-1434541323-headset.flac", + "question": "tell me the first word that's in relation to 'house_place' in this audio?", + "choice_a": "living", + "choice_b": "the", + "choice_c": "living room", + "choice_d": "bedroom", + "answer_gt": "living room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7563 + }, + { + "path": "audio-1498568872.flac", + "question": "Which word that's affiliated with 'time' is spoken first in this clip?", + "choice_a": "have", + "choice_b": "do", + "choice_c": "six am", + "choice_d": "past three months", + "answer_gt": "six am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7564 + }, + { + "path": "audio-1501691801-headset.flac", + "question": "name the first word that's related to 'house_place' in this audio?", + "choice_a": "light", + "choice_b": "garage", + "choice_c": "house", + "choice_d": "my son's room", + "answer_gt": "garage", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7565 + }, + { + "path": "audio-1501690611.flac", + "question": "What's the first word that's indicative of 'timeofday' in this audio?", + "choice_a": "my", + "choice_b": "schedule", + "choice_c": "afternoon", + "choice_d": "this afternoon", + "answer_gt": "afternoon", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7566 + }, + { + "path": "audio-1499267014.flac", + "question": "What's the first word in this audio that has a connection to 'food_type'?", + "choice_a": "how", + "choice_b": "i", + "choice_c": "stuffed peppers", + "choice_d": "risotto", + "answer_gt": "stuffed peppers", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7567 + }, + { + "path": "audio-1501754037-headset.flac", + "question": "What's the premier word about 'media_type' in this sound clip?", + "choice_a": "for", + "choice_b": "that", + "choice_c": "podcasts", + "choice_d": "audio book", + "answer_gt": "podcasts", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7568 + }, + { + "path": "audio-1498574489.flac", + "question": "What is the first word that's in connection with 'date' in this audio?", + "choice_a": "have", + "choice_b": "a", + "choice_c": "friday", + "choice_d": "march eighteen", + "answer_gt": "friday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7569 + }, + { + "path": "audio-1501690612.flac", + "question": "Which word comes first in this audio that's tied to 'list_name'?", + "choice_a": "my", + "choice_b": "from", + "choice_c": "old english", + "choice_d": "expenses", + "answer_gt": "old english", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7570 + }, + { + "path": "audio-1501751396-headset.flac", + "question": "Which word linked to 'place_name' comes first in this recording?", + "choice_a": "condition", + "choice_b": "my place", + "choice_c": "beijing", + "choice_d": "london", + "answer_gt": "my place", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7571 + }, + { + "path": "audio--1505407237-headset.flac", + "question": "name the first word that's related to 'player_setting' in this audio?", + "choice_a": "audiobook", + "choice_b": "resume", + "choice_c": "playback", + "choice_d": "move on", + "answer_gt": "resume", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7572 + }, + { + "path": "audio-1496266208.flac", + "question": "What's the first word that's alluding to 'place_name' in this sound recording?", + "choice_a": "me", + "choice_b": "tell", + "choice_c": "jaipur", + "choice_d": "kazakhstan", + "answer_gt": "jaipur", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7573 + }, + { + "path": "audio-1490802505-headset.flac", + "question": "Which word that signifies 'weather_descriptor' is spoken first in this clip?", + "choice_a": "i", + "choice_b": "my", + "choice_c": "raining", + "choice_d": "sandals", + "answer_gt": "sandals", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7574 + }, + { + "path": "audio-1501765068.flac", + "question": "tell me the first word that's in relation to 'news_topic' in this audio?", + "choice_a": "set", + "choice_b": "environmental", + "choice_c": "political", + "choice_d": "new york bridge proposal", + "answer_gt": "environmental", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7575 + }, + { + "path": "audio-1498486635-headset.flac", + "question": "What's the first word that's representative of 'date' in this recording?", + "choice_a": "today's", + "choice_b": "date", + "choice_c": "second march", + "choice_d": "one thousand nine hundred and ninety", + "answer_gt": "today's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7576 + }, + { + "path": "audio--1504195056-headset.flac", + "question": "What's the first word that's indicative of 'time' in this audio?", + "choice_a": "eight am", + "choice_b": "turn", + "choice_c": "four", + "choice_d": "midnight", + "answer_gt": "eight am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7577 + }, + { + "path": "audio-1488562775.flac", + "question": "What's the first word that's associated with 'event_name' in this clip?", + "choice_a": "fringe", + "choice_b": "next", + "choice_c": "eric's bday", + "choice_d": "easter", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7578 + }, + { + "path": "audio-1501698447.flac", + "question": "What is the initial 'list_name'-associated word you hear in this recording?", + "choice_a": "walmart list", + "choice_b": "the", + "choice_c": "contact list", + "choice_d": "old music", + "answer_gt": "walmart list", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7579 + }, + { + "path": "audio-1497439869.flac", + "question": "identify the initial word associated with 'podcast_descriptor' in this audio.", + "choice_a": "i", + "choice_b": "started last night", + "choice_c": "next podcast", + "choice_d": "rock songs having good rating", + "answer_gt": "started last night", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7580 + }, + { + "path": "audio-1497886122.flac", + "question": "Which word comes first in this audio that's tied to 'device_type'?", + "choice_a": "my", + "choice_b": "disable", + "choice_c": "okug", + "choice_d": "smart plug", + "answer_gt": "okug", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7581 + }, + { + "path": "audio-1434541337.flac", + "question": "What is the first word that's in connection with 'business_name' in this audio?", + "choice_a": "nearest", + "choice_b": "the", + "choice_c": "walmart", + "choice_d": "googs", + "answer_gt": "walmart", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7582 + }, + { + "path": "audio-1502195789-headset.flac", + "question": "Which word that's linked to 'event_name' do you hear first in this recording?", + "choice_a": "up", + "choice_b": "pick up my parents", + "choice_c": "gun shows", + "choice_d": "easter", + "answer_gt": "pick up my parents", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7583 + }, + { + "path": "audio-1501685995-headset.flac", + "question": "What's the first word that's alluding to 'time' in this sound recording?", + "choice_a": "new", + "choice_b": "am", + "choice_c": "three thirty pm", + "choice_d": "eight am", + "answer_gt": "eight am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7584 + }, + { + "path": "audio-1498573203-headset.flac", + "question": "What's the first word that's representative of 'media_type' in this recording?", + "choice_a": "news", + "choice_b": "from", + "choice_c": "social networks", + "choice_d": "cnn", + "answer_gt": "cnn", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7585 + }, + { + "path": "audio-1490200738-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'song_name'?", + "choice_a": "i", + "choice_b": "want", + "choice_c": "purple haze", + "choice_d": "thirteen", + "answer_gt": "purple haze", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7586 + }, + { + "path": "audio-1494512985-headset.flac", + "question": "What word relevant to 'date' is mentioned at the start of this audio?", + "choice_a": "feeling", + "choice_b": "you", + "choice_c": "today", + "choice_d": "twenty first", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7587 + }, + { + "path": "audio-1498567668.flac", + "question": "tell me the first word that's in relation to 'device_type' in this audio?", + "choice_a": "the", + "choice_b": "start", + "choice_c": "coffee pot", + "choice_d": "vacuum cleaner", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7588 + }, + { + "path": "audio-1490108335-headset.flac", + "question": "What begins the list of 'place_name' words in this audio clip?", + "choice_a": "was", + "choice_b": "when", + "choice_c": "everest", + "choice_d": "us", + "answer_gt": "us", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7589 + }, + { + "path": "audio-1497863317.flac", + "question": "What's the first word that's in reference to 'relation' in this sound clip?", + "choice_a": "email", + "choice_b": "send", + "choice_c": "mom's", + "choice_d": "mom", + "answer_gt": "mom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7590 + }, + { + "path": "audio-1502218342-headset.flac", + "question": "What begins the list of 'event_name' words in this audio clip?", + "choice_a": "a", + "choice_b": "sports game", + "choice_c": "pick up my parents", + "choice_d": "pick up my dry cleaning", + "answer_gt": "sports game", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7591 + }, + { + "path": "audio-1488561845.flac", + "question": "Which word that's affiliated with 'house_place' is spoken first in this clip?", + "choice_a": "turn", + "choice_b": "in", + "choice_c": "room", + "choice_d": "bedroom", + "answer_gt": "bedroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7592 + }, + { + "path": "audio--1504190285-headset.flac", + "question": "What's the first word that's characteristic of 'podcast_name' in this clip?", + "choice_a": "young turks", + "choice_b": "play", + "choice_c": "order of the green hand", + "choice_d": "disney", + "answer_gt": "young turks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7593 + }, + { + "path": "audio-1501758730-headset.flac", + "question": "What's the first word that's in reference to 'event_name' in this sound clip?", + "choice_a": "my", + "choice_b": "from", + "choice_c": "meetings", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7594 + }, + { + "path": "audio-1501693645-headset.flac", + "question": "What is the first word that's in connection with 'event_name' in this audio?", + "choice_a": "theme parks", + "choice_b": "what", + "choice_c": "easter", + "choice_d": "supper", + "answer_gt": "theme parks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7595 + }, + { + "path": "audio-1494608777-headset.flac", + "question": "Which word that's linked to 'game_name' do you hear first in this recording?", + "choice_a": "play", + "choice_b": "lets", + "choice_c": "contra", + "choice_d": "battleship", + "answer_gt": "battleship", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7596 + }, + { + "path": "audio-1497861959.flac", + "question": "Which word is introduced first in this audio that relates to 'person'?", + "choice_a": "joanna", + "choice_b": "set", + "choice_c": "naty", + "choice_d": "bob", + "answer_gt": "joanna", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7597 + }, + { + "path": "audio-1499265711-headset.flac", + "question": "Which term that relates to 'event_name' is spoken first in this clip?", + "choice_a": "appointment", + "choice_b": "remove", + "choice_c": "pre season baseball", + "choice_d": "parade", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7598 + }, + { + "path": "audio-1498570575-headset.flac", + "question": "Which term that relates to 'weather_descriptor' is spoken first in this clip?", + "choice_a": "this", + "choice_b": "will", + "choice_c": "rain", + "choice_d": "sweater", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7599 + }, + { + "path": "audio-1488538753.flac", + "question": "What is the first word related to 'house_place' mentioned in this audio clip?", + "choice_a": "in", + "choice_b": "the", + "choice_c": "bedroom", + "choice_d": "house", + "answer_gt": "bedroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7600 + }, + { + "path": "audio-1501687054-headset.flac", + "question": "What is the first word that's in connection with 'event_name' in this audio?", + "choice_a": "doctor appointment", + "choice_b": "what", + "choice_c": "lunch meeting", + "choice_d": "musical events", + "answer_gt": "doctor appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7601 + }, + { + "path": "audio-1490007028-headset.flac", + "question": "Which term that relates to 'meal_type' is spoken first in this clip?", + "choice_a": "ready", + "choice_b": "is", + "choice_c": "dinner", + "choice_d": "lunch", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7602 + }, + { + "path": "audio-1494548467-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'weather_descriptor'?", + "choice_a": "rainy", + "choice_b": "be", + "choice_c": "precipitation", + "choice_d": "raining", + "answer_gt": "rainy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7603 + }, + { + "path": "audio-1497869637-headset.flac", + "question": "Which word is the earliest mention of something related to 'transport_type' in this audio?", + "choice_a": "train", + "choice_b": "train", + "choice_c": "cab", + "choice_d": "yellow cab", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7604 + }, + { + "path": "audio-1494415805.flac", + "question": "Which term that relates to 'device_type' is spoken first in this clip?", + "choice_a": "how", + "choice_b": "vacuum", + "choice_c": "robot vacuum cleaner", + "choice_d": "lamps", + "answer_gt": "robot vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7605 + }, + { + "path": "audio-1490357777-headset.flac", + "question": "Which term that relates to 'date' is spoken first in this clip?", + "choice_a": "is", + "choice_b": "weather", + "choice_c": "today's", + "choice_d": "friday", + "answer_gt": "friday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7606 + }, + { + "path": "audio-1497882400.flac", + "question": "Which word is introduced first in this audio that relates to 'person'?", + "choice_a": "seth", + "choice_b": "where", + "choice_c": "george", + "choice_d": "morgan freeman", + "answer_gt": "seth", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7607 + }, + { + "path": "audio-1490703414-headset.flac", + "question": "Which word is the first to be related to 'transport_type' in this recording?", + "choice_a": "is", + "choice_b": "train", + "choice_c": "uberpool", + "choice_d": "taxi", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7608 + }, + { + "path": "audio-1501250793-headset.flac", + "question": "What's the first word that's tied to 'definition_word' mentioned in this audio?", + "choice_a": "what", + "choice_b": "is", + "choice_c": "turbulence", + "choice_d": "caftan", + "answer_gt": "caftan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7609 + }, + { + "path": "audio-1501688280.flac", + "question": "What's the first word that comes up in this audio in relation to 'device_type'?", + "choice_a": "the", + "choice_b": "smart", + "choice_c": "smart socket", + "choice_d": "smart plug", + "answer_gt": "smart plug", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7610 + }, + { + "path": "audio-1494422717.flac", + "question": "What's the first word that comes up in this audio in relation to 'podcast_name'?", + "choice_a": "from", + "choice_b": "green", + "choice_c": "young turks", + "choice_d": "order of the green hand", + "answer_gt": "order of the green hand", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7611 + }, + { + "path": "audio-1489673960-headset.flac", + "question": "Which word comes first in this audio that's tied to 'device_type'?", + "choice_a": "vacuum cleaner", + "choice_b": "ten", + "choice_c": "plug", + "choice_d": "robot vacuum cleaner", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7612 + }, + { + "path": "audio-1490264244-headset.flac", + "question": "Which word that's evocative of 'event_name' is mentioned first in this audio?", + "choice_a": "pick up my dry cleaning", + "choice_b": "up", + "choice_c": "medical appointment", + "choice_d": "call mom", + "answer_gt": "pick up my dry cleaning", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7613 + }, + { + "path": "audio--1504192976.flac", + "question": "Which term that relates to 'player_setting' is spoken first in this clip?", + "choice_a": "keep", + "choice_b": "secret", + "choice_c": "keep playing", + "choice_d": "pause", + "answer_gt": "keep playing", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7614 + }, + { + "path": "audio-1434541952-headset.flac", + "question": "Which word linked to 'news_topic' comes first in this recording?", + "choice_a": "tornadoes", + "choice_b": "many", + "choice_c": "politics", + "choice_d": "weather", + "answer_gt": "tornadoes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7615 + }, + { + "path": "audio-1501775422.flac", + "question": "What is the first word in this clip that's relevant to 'currency_name'?", + "choice_a": "the", + "choice_b": "us dollar", + "choice_c": "u. s. d.", + "choice_d": "euro", + "answer_gt": "us dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7616 + }, + { + "path": "audio-1501767923-headset.flac", + "question": "Which word is the first to be related to 'player_setting' in this recording?", + "choice_a": "to", + "choice_b": "change", + "choice_c": "jump", + "choice_d": "hold", + "answer_gt": "change", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7617 + }, + { + "path": "audio-1490369148-headset.flac", + "question": "What word relevant to 'list_name' is mentioned at the start of this audio?", + "choice_a": "delete", + "choice_b": "do's", + "choice_c": "to do's", + "choice_d": "activities", + "answer_gt": "to do's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7618 + }, + { + "path": "audio-1494422311.flac", + "question": "What word relevant to 'date' is mentioned at the start of this audio?", + "choice_a": "about", + "choice_b": "a", + "choice_c": "tomorrow's", + "choice_d": "a day before", + "answer_gt": "tomorrow's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7619 + }, + { + "path": "audio-1489152284.flac", + "question": "What's the first word that comes up in this audio in relation to 'person'?", + "choice_a": "email", + "choice_b": "please", + "choice_c": "lisa ann", + "choice_d": "jeff", + "answer_gt": "jeff", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7620 + }, + { + "path": "audio-1490106723-headset.flac", + "question": "What's the first word that's representative of 'playlist_name' in this recording?", + "choice_a": "summer jams playlist", + "choice_b": "jams", + "choice_c": "itunes library", + "choice_d": "running", + "answer_gt": "summer jams playlist", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7621 + }, + { + "path": "audio-1502892988-headset.flac", + "question": "What's the first word that's in association with 'date' in this recording?", + "choice_a": "received", + "choice_b": "today", + "choice_c": "weekdays", + "choice_d": "this week's", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7622 + }, + { + "path": "audio-1501687038.flac", + "question": "What's the first word that's in association with 'business_name' in this recording?", + "choice_a": "louies", + "choice_b": "take", + "choice_c": "company name", + "choice_d": "hdfc", + "answer_gt": "louies", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7623 + }, + { + "path": "audio-1502112632-headset.flac", + "question": "What's the first word that's tied to 'music_genre' mentioned in this audio?", + "choice_a": "all", + "choice_b": "pop", + "choice_c": "smooth jazz", + "choice_d": "sad songs", + "answer_gt": "pop", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7624 + }, + { + "path": "audio-1499690358.flac", + "question": "What word relevant to 'place_name' is mentioned at the start of this audio?", + "choice_a": "london", + "choice_b": "is", + "choice_c": "west virginia", + "choice_d": "india's", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7625 + }, + { + "path": "audio-1490967578.flac", + "question": "What is the first word related to 'event_name' mentioned in this audio clip?", + "choice_a": "accounting", + "choice_b": "meeting", + "choice_c": "musical events", + "choice_d": "pick up my parents", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7626 + }, + { + "path": "audio-1501688719.flac", + "question": "What is the first word that's associated with 'media_type' in this audio?", + "choice_a": "that", + "choice_b": "open", + "choice_c": "tweet", + "choice_d": "newest provider", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7627 + }, + { + "path": "audio-1499246638.flac", + "question": "find the first word that's linked to 'color_type' in this sound clip?", + "choice_a": "is", + "choice_b": "pink", + "choice_c": "blue", + "choice_d": "red", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7628 + }, + { + "path": "audio-1490368192.flac", + "question": "tell me the first 'music_genre'-connected word in this audio.", + "choice_a": "play", + "choice_b": "reggae", + "choice_c": "smooth jazz", + "choice_d": "rock", + "answer_gt": "reggae", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7629 + }, + { + "path": "audio-1502113575.flac", + "question": "tell me the first word that's in relation to 'artist_name' in this audio?", + "choice_a": "doom", + "choice_b": "play", + "choice_c": "mf doom", + "choice_d": "david bowie", + "answer_gt": "mf doom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7630 + }, + { + "path": "audio-1490199175.flac", + "question": "Which word is the first to be related to 'change_amount' in this recording?", + "choice_a": "volume", + "choice_b": "by fifty percent", + "choice_c": "twenty percent", + "choice_d": "to maximum", + "answer_gt": "by fifty percent", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7631 + }, + { + "path": "audio-1502198990.flac", + "question": "What's the first word that's in association with 'event_name' in this recording?", + "choice_a": "one", + "choice_b": "set", + "choice_c": "lunch meeting", + "choice_d": "wedding", + "answer_gt": "lunch meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7632 + }, + { + "path": "audio-1494512463-headset.flac", + "question": "What's the first word that's alluding to 'music_descriptor' in this sound recording?", + "choice_a": "play", + "choice_b": "from", + "choice_c": "good song", + "choice_d": "sad", + "answer_gt": "good song", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7633 + }, + { + "path": "audio-1502197310-headset.flac", + "question": "Which word linked to 'date' comes first in this recording?", + "choice_a": "whats", + "choice_b": "weather", + "choice_c": "tomorrow", + "choice_d": "everyday", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7634 + }, + { + "path": "audio-1499688834-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'business_type'?", + "choice_a": "cinema", + "choice_b": "movies", + "choice_c": "furniture store", + "choice_d": "stores", + "answer_gt": "cinema", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7635 + }, + { + "path": "audio-1501686116.flac", + "question": "Which word that's linked to 'transport_type' do you hear first in this recording?", + "choice_a": "a", + "choice_b": "taxi", + "choice_c": "cab", + "choice_d": "train", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7636 + }, + { + "path": "audio-1502201226-headset.flac", + "question": "What is the opening word pertaining to 'time' mentioned here?", + "choice_a": "am", + "choice_b": "manila", + "choice_c": "one hour", + "choice_d": "eight am", + "answer_gt": "eight am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7637 + }, + { + "path": "audio--1504192261-headset.flac", + "question": "What's the earliest 'player_setting'-related term you hear in this clip?", + "choice_a": "stop", + "choice_b": "stop", + "choice_c": "stop play", + "choice_d": "start chapter five", + "answer_gt": "stop", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7638 + }, + { + "path": "audio-1494418756-headset.flac", + "question": "In this sound clip, which word related to 'date' is mentioned first?", + "choice_a": "off", + "choice_b": "the", + "choice_c": "rest of the year", + "choice_d": "monday", + "answer_gt": "rest of the year", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7639 + }, + { + "path": "audio-1490202467-headset.flac", + "question": "What begins the list of 'news_topic' words in this audio clip?", + "choice_a": "give", + "choice_b": "president trump", + "choice_c": "politics", + "choice_d": "cars", + "answer_gt": "president trump", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7640 + }, + { + "path": "audio-1502216136-headset.flac", + "question": "tell me the first word that's in relation to 'time' in this audio?", + "choice_a": "eight am", + "choice_b": "eight", + "choice_c": "another hour", + "choice_d": "midnight", + "answer_gt": "eight am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7641 + }, + { + "path": "audio-1489494280-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'cooking_type'?", + "choice_a": "how", + "choice_b": "boil", + "choice_c": "boiling", + "choice_d": "boil", + "answer_gt": "boil", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7642 + }, + { + "path": "audio-1488992754.flac", + "question": "What's the first word that's reflective of 'podcast_name' in this sound clip?", + "choice_a": "disney", + "choice_b": "find", + "choice_c": "mike and mike", + "choice_d": "order of the green hand", + "answer_gt": "disney", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7643 + }, + { + "path": "audio-1501777061-headset.flac", + "question": "In this audio, what's the first word you hear that's connected to 'place_name'?", + "choice_a": "which", + "choice_b": "peak", + "choice_c": "town", + "choice_d": "world", + "answer_gt": "world", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7644 + }, + { + "path": "audio-1490203529-headset.flac", + "question": "What's the earliest 'house_place'-related term you hear in this clip?", + "choice_a": "room's", + "choice_b": "the", + "choice_c": "patio", + "choice_d": "living room's", + "answer_gt": "living room's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7645 + }, + { + "path": "audio-1490202252-headset.flac", + "question": "Which word that's evocative of 'email_folder' is mentioned first in this audio?", + "choice_a": "new", + "choice_b": "inform", + "choice_c": "my", + "choice_d": "inbox", + "answer_gt": "inbox", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7646 + }, + { + "path": "audio-1497263161-headset.flac", + "question": "detect the first mention of a word related to 'date' in this audio.", + "choice_a": "weather", + "choice_b": "the", + "choice_c": "this week", + "choice_d": "rest of the year", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7647 + }, + { + "path": "audio-1490282978-headset.flac", + "question": "What is the initial 'event_name'-associated word you hear in this recording?", + "choice_a": "eight", + "choice_b": "to", + "choice_c": "pick up my dry cleaning", + "choice_d": "meetings", + "answer_gt": "pick up my dry cleaning", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7648 + }, + { + "path": "audio-1490108508.flac", + "question": "find the first word that's linked to 'event_name' in this sound clip?", + "choice_a": "to", + "choice_b": "birthday", + "choice_c": "soccer match", + "choice_d": "lee's birthday", + "answer_gt": "birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7649 + }, + { + "path": "audio-1502308898-headset.flac", + "question": "What is the first word in this clip that's relevant to 'relation'?", + "choice_a": "birthday", + "choice_b": "a", + "choice_c": "mom's", + "choice_d": "mother's", + "answer_gt": "mom's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7650 + }, + { + "path": "audio-1501688646.flac", + "question": "What's the first word that's associated with 'date' in this clip?", + "choice_a": "like", + "choice_b": "the", + "choice_c": "today", + "choice_d": "friday", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7651 + }, + { + "path": "audio--1506079542-headset.flac", + "question": "Which word that signifies 'audiobook_name' is spoken first in this clip?", + "choice_a": "you", + "choice_b": "tarzan", + "choice_c": "pride and prejudice", + "choice_d": "hunt for read october", + "answer_gt": "tarzan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7652 + }, + { + "path": "audio-1497884886-headset.flac", + "question": "What's the first word that's connected to 'device_type' in this audio?", + "choice_a": "enable", + "choice_b": "my", + "choice_c": "plug", + "choice_d": "robot vacuum cleaner", + "answer_gt": "plug", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7653 + }, + { + "path": "audio-1501696009-headset.flac", + "question": "Which word is the first to be related to 'music_descriptor' in this recording?", + "choice_a": "play", + "choice_b": "previous", + "choice_c": "sad", + "choice_d": "previous", + "answer_gt": "previous", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7654 + }, + { + "path": "audio-1490006168.flac", + "question": "What is the first word in this clip that's relevant to 'definition_word'?", + "choice_a": "hurricane", + "choice_b": "all", + "choice_c": "turbulence", + "choice_d": "book", + "answer_gt": "hurricane", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7655 + }, + { + "path": "audio-1501157457-headset.flac", + "question": "What word relevant to 'podcast_name' is mentioned at the start of this audio?", + "choice_a": "i", + "choice_b": "mike and mike", + "choice_c": "order of the green hand", + "choice_d": "sports podcast", + "answer_gt": "mike and mike", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7656 + }, + { + "path": "audio--1505405859.flac", + "question": "What's the earliest 'player_setting'-related term you hear in this clip?", + "choice_a": "the", + "choice_b": "read", + "choice_c": "continue", + "choice_d": "playback", + "answer_gt": "read", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7657 + }, + { + "path": "audio-1490108297-headset.flac", + "question": "What is the opening word pertaining to 'house_place' mentioned here?", + "choice_a": "in", + "choice_b": "lower", + "choice_c": "kitchen", + "choice_d": "floor", + "answer_gt": "kitchen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7658 + }, + { + "path": "audio--1504192948-headset.flac", + "question": "What's the first word that's suggestive of 'alarm_type' in this audio recording?", + "choice_a": "up", + "choice_b": "this", + "choice_c": "wake up", + "choice_d": "soccer practice", + "answer_gt": "wake up", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7659 + }, + { + "path": "audio-1499245963-headset.flac", + "question": "What's the earliest 'relation'-related term you hear in this clip?", + "choice_a": "i", + "choice_b": "dad", + "choice_c": "friend", + "choice_d": "mom's", + "answer_gt": "dad", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7660 + }, + { + "path": "audio-1490106229.flac", + "question": "What is the initial 'place_name'-associated word you hear in this recording?", + "choice_a": "nigeria", + "choice_b": "is", + "choice_c": "the equator", + "choice_d": "alaska", + "answer_gt": "nigeria", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7661 + }, + { + "path": "audio-1502200746-headset.flac", + "question": "Which word that's linked to 'time_zone' do you hear first in this recording?", + "choice_a": "what", + "choice_b": "eastern", + "choice_c": "eastern", + "choice_d": "pacific", + "answer_gt": "eastern", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7662 + }, + { + "path": "audio-1490199616.flac", + "question": "In this audio, what's the first word you hear that's connected to 'player_setting'?", + "choice_a": "pause", + "choice_b": "the", + "choice_c": "start chapter five", + "choice_d": "resume", + "answer_gt": "pause", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7663 + }, + { + "path": "audio-1490359317-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'app_name'?", + "choice_a": "pandora", + "choice_b": "start", + "choice_c": "start", + "choice_d": "start", + "answer_gt": "pandora", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7664 + }, + { + "path": "audio-1502101259.flac", + "question": "What's the first word mentioned in this audio that's about 'order_type'?", + "choice_a": "they", + "choice_b": "can", + "choice_c": "takeout", + "choice_d": "takeaway", + "answer_gt": "takeaway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7665 + }, + { + "path": "audio-1490286754.flac", + "question": "Which word is introduced first in this audio that relates to 'device_type'?", + "choice_a": "off", + "choice_b": "the", + "choice_c": "lamps", + "choice_d": "robot roomba", + "answer_gt": "lamps", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7666 + }, + { + "path": "audio-1501691822-headset.flac", + "question": "What word is the first to reference 'date' in this sound clip?", + "choice_a": "of", + "choice_b": "two days", + "choice_c": "tomorrow", + "choice_d": "Friday", + "answer_gt": "two days", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7667 + }, + { + "path": "audio-1489497748.flac", + "question": "find the first word that's linked to 'artist_name' in this sound clip?", + "choice_a": "play", + "choice_b": "slayer", + "choice_c": "justin bieber's", + "choice_d": "joe prsaise", + "answer_gt": "slayer", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7668 + }, + { + "path": "audio-1501778782.flac", + "question": "What is the first word related to 'playlist_name' mentioned in this audio clip?", + "choice_a": "to", + "choice_b": "add", + "choice_c": "rock list", + "choice_d": "running", + "answer_gt": "running", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7669 + }, + { + "path": "audio-1490261991-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'person'?", + "choice_a": "was", + "choice_b": "elvis presley's", + "choice_c": "jack", + "choice_d": "mary's", + "answer_gt": "elvis presley's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7670 + }, + { + "path": "audio-1498564646.flac", + "question": "find the first word that's linked to 'business_name' in this sound clip?", + "choice_a": "pizza hut", + "choice_b": "hut", + "choice_c": "charley's steakhouse", + "choice_d": "samsung", + "answer_gt": "pizza hut", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7671 + }, + { + "path": "audio-1490703567.flac", + "question": "In this audio clip, what's the first word that pertains to 'song_name'?", + "choice_a": "to", + "choice_b": "bohemian", + "choice_c": "jingle bells", + "choice_d": "bohemian raphsody", + "answer_gt": "bohemian raphsody", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7672 + }, + { + "path": "audio-1502896000-headset.flac", + "question": "What is the first word in this clip that's relevant to 'food_type'?", + "choice_a": "make", + "choice_b": "noodles", + "choice_c": "tuna", + "choice_d": "meatballs", + "answer_gt": "noodles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7673 + }, + { + "path": "audio-1497860864.flac", + "question": "name the first word that's related to 'currency_name' in this audio?", + "choice_a": "british pound", + "choice_b": "british", + "choice_c": "dollar", + "choice_d": "pounds", + "answer_gt": "british pound", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7674 + }, + { + "path": "audio-1497867008-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'date'?", + "choice_a": "what", + "choice_b": "tuesday", + "choice_c": "tomorrow's", + "choice_d": "this weekend", + "answer_gt": "tuesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7675 + }, + { + "path": "audio-1500037176.flac", + "question": "What is the first word in this clip that's relevant to 'transport_type'?", + "choice_a": "train", + "choice_b": "times", + "choice_c": "metro", + "choice_d": "uberpool", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7676 + }, + { + "path": "audio-1502375944-headset.flac", + "question": "Which word comes first in this audio that's tied to 'place_name'?", + "choice_a": "active", + "choice_b": "earth", + "choice_c": "town", + "choice_d": "berlin", + "answer_gt": "earth", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7677 + }, + { + "path": "audio-1502377573.flac", + "question": "tell me the first 'device_type'-connected word in this audio.", + "choice_a": "vacuum cleaner", + "choice_b": "turn", + "choice_c": "robot roomba", + "choice_d": "smart socket", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7678 + }, + { + "path": "audio-1434536404.flac", + "question": "What's the premier word about 'news_topic' in this sound clip?", + "choice_a": "me", + "choice_b": "bitcoin price", + "choice_c": "the brexit", + "choice_d": "environmental", + "answer_gt": "bitcoin price", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7679 + }, + { + "path": "audio-1501691380-headset.flac", + "question": "Which word is the first to be related to 'time' in this recording?", + "choice_a": "is", + "choice_b": "for", + "choice_c": "ten", + "choice_d": "ten am", + "answer_gt": "ten am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7680 + }, + { + "path": "audio-1501406812-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'music_genre'?", + "choice_a": "rock", + "choice_b": "the", + "choice_c": "pop", + "choice_d": "techno", + "answer_gt": "rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7681 + }, + { + "path": "audio-1501751913.flac", + "question": "What begins the list of 'place_name' words in this audio clip?", + "choice_a": "theaters", + "choice_b": "doesn't", + "choice_c": "location", + "choice_d": "india's", + "answer_gt": "theaters", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7682 + }, + { + "path": "audio-1497557781.flac", + "question": "name the first word that's related to 'definition_word' in this audio?", + "choice_a": "the", + "choice_b": "is", + "choice_c": "logic", + "choice_d": "orange", + "answer_gt": "logic", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7683 + }, + { + "path": "audio-1490103572.flac", + "question": "What's the first word that comes up in this audio in relation to 'music_genre'?", + "choice_a": "my", + "choice_b": "play", + "choice_c": "rock", + "choice_d": "pop", + "answer_gt": "rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7684 + }, + { + "path": "audio-1501151973.flac", + "question": "What's the first term that pertains to 'time' in this recording?", + "choice_a": "to", + "choice_b": "a", + "choice_c": "three thirty pm", + "choice_d": "a quarter to two", + "answer_gt": "a quarter to two", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7685 + }, + { + "path": "audio-1490261671-headset.flac", + "question": "identify the initial word associated with 'date' in this audio.", + "choice_a": "today", + "choice_b": "is", + "choice_c": "monday", + "choice_d": "this weekend", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7686 + }, + { + "path": "audio-1490106508.flac", + "question": "In this audio clip, what's the first word that pertains to 'color_type'?", + "choice_a": "the", + "choice_b": "pink", + "choice_c": "red", + "choice_d": "blue", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7687 + }, + { + "path": "audio-1502894668.flac", + "question": "What's the first word that's associated with 'transport_type' in this clip?", + "choice_a": "a", + "choice_b": "ticket", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7688 + }, + { + "path": "audio-1501752961-headset.flac", + "question": "What's the first word that's characteristic of 'place_name' in this clip?", + "choice_a": "tell", + "choice_b": "grand rapids mi", + "choice_c": "disney land", + "choice_d": "bolshoi theatre", + "answer_gt": "grand rapids mi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7689 + }, + { + "path": "audio-1497022798.flac", + "question": "What's the first word in this audio that has a connection to 'device_type'?", + "choice_a": "start", + "choice_b": "vacuum", + "choice_c": "all the lights", + "choice_d": "vacuum cleaner", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7690 + }, + { + "path": "audio-1499690922.flac", + "question": "In this audio, what's the first word that's reminiscent of 'definition_word'?", + "choice_a": "me", + "choice_b": "the", + "choice_c": "smartphone circuit", + "choice_d": "television circuit", + "answer_gt": "smartphone circuit", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7691 + }, + { + "path": "audio-1497029450.flac", + "question": "What's the first word mentioned in this audio that's about 'house_place'?", + "choice_a": "light", + "choice_b": "make", + "choice_c": "living room's", + "choice_d": "room", + "answer_gt": "room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7692 + }, + { + "path": "audio-1490106212.flac", + "question": "What's the first word that's connected to 'media_type' in this audio?", + "choice_a": "social networks", + "choice_b": "social", + "choice_c": "facebook", + "choice_d": "time of india", + "answer_gt": "social networks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7693 + }, + { + "path": "audio-1490107162.flac", + "question": "What's the first word that's in association with 'person' in this recording?", + "choice_a": "jerry", + "choice_b": "email", + "choice_c": "jeff", + "choice_d": "chelsea", + "answer_gt": "jerry", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7694 + }, + { + "path": "audio-1494418564-headset.flac", + "question": "What's the first word that's suggestive of 'date' in this audio recording?", + "choice_a": "only", + "choice_b": "six", + "choice_c": "today's", + "choice_d": "monday", + "answer_gt": "monday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7695 + }, + { + "path": "audio-1497884787.flac", + "question": "What is the first word related to 'date' mentioned in this audio clip?", + "choice_a": "the", + "choice_b": "for", + "choice_c": "twenty first", + "choice_d": "week", + "answer_gt": "twenty first", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7696 + }, + { + "path": "audio-1490262348.flac", + "question": "In this sound clip, which word related to 'list_name' is mentioned first?", + "choice_a": "to", + "choice_b": "add", + "choice_c": "grocery", + "choice_d": "party time", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7697 + }, + { + "path": "audio-1497880065-headset.flac", + "question": "What's the first word in this audio that has a connection to 'person'?", + "choice_a": "dot", + "choice_b": "contacts", + "choice_c": "john doe", + "choice_d": "elvis presley's", + "answer_gt": "john doe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7698 + }, + { + "path": "audio--1504192272.flac", + "question": "What's the first word that's tied to 'alarm_type' mentioned in this audio?", + "choice_a": "please", + "choice_b": "at", + "choice_c": "soccer practice", + "choice_d": "wake up", + "answer_gt": "wake up", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7699 + }, + { + "path": "audio-1501778696-headset.flac", + "question": "Which term that relates to 'event_name' is spoken first in this clip?", + "choice_a": "about", + "choice_b": "the", + "choice_c": "party", + "choice_d": "haircut", + "answer_gt": "party", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7700 + }, + { + "path": "audio-1489153354.flac", + "question": "Which word that's evocative of 'house_place' is mentioned first in this audio?", + "choice_a": "office", + "choice_b": "lights", + "choice_c": "bedroom", + "choice_d": "floor", + "answer_gt": "office", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7701 + }, + { + "path": "audio-1434529698-headset.flac", + "question": "Which word that's linked to 'device_type' do you hear first in this recording?", + "choice_a": "the", + "choice_b": "coffee machine", + "choice_c": "vacuum cleaner", + "choice_d": "vacuum", + "answer_gt": "coffee machine", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7702 + }, + { + "path": "audio-1490008019-headset.flac", + "question": "What's the first word that's reflective of 'media_type' in this sound clip?", + "choice_a": "has", + "choice_b": "me", + "choice_c": "social networks", + "choice_d": "twitter", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7703 + }, + { + "path": "audio--1505404392-headset.flac", + "question": "spot the first word that hints at 'business_type' in this audio.", + "choice_a": "local", + "choice_b": "pubs", + "choice_c": "stores", + "choice_d": "furniture store", + "answer_gt": "pubs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7704 + }, + { + "path": "audio-1498565442.flac", + "question": "detect the first mention of a word related to 'weather_descriptor' in this audio.", + "choice_a": "any", + "choice_b": "rain or snow", + "choice_c": "nice", + "choice_d": "rainy", + "answer_gt": "rain or snow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7705 + }, + { + "path": "audio-1490261585-headset.flac", + "question": "Which word that's evocative of 'coffee_type' is mentioned first in this audio?", + "choice_a": "hey", + "choice_b": "me", + "choice_c": "creamy", + "choice_d": "salted carmel flavored cream", + "answer_gt": "creamy", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7706 + }, + { + "path": "audio-1502216638.flac", + "question": "point out the first 'place_name'-related word uttered in this audio.", + "choice_a": "russia", + "choice_b": "from", + "choice_c": "la", + "choice_d": "austin", + "answer_gt": "russia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7707 + }, + { + "path": "audio-1501686579.flac", + "question": "Which word that's affiliated with 'transport_type' is spoken first in this clip?", + "choice_a": "book", + "choice_b": "in", + "choice_c": "taxi", + "choice_d": "metro", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7708 + }, + { + "path": "audio-1501690561.flac", + "question": "What's the first word that comes up in this audio in relation to 'date'?", + "choice_a": "today", + "choice_b": "so", + "choice_c": "rest of the year", + "choice_d": "monday the twentieth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7709 + }, + { + "path": "audio-1490291360.flac", + "question": "What is the first word that's in connection with 'currency_name' in this audio?", + "choice_a": "of", + "choice_b": "rate", + "choice_c": "euro", + "choice_d": "u. s. d.", + "answer_gt": "euro", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7710 + }, + { + "path": "audio-1502891175.flac", + "question": "Which word that's synonymous with 'person' appears first in this audio?", + "choice_a": "from", + "choice_b": "jeffrey burnette", + "choice_c": "cassy", + "choice_d": "obama", + "answer_gt": "jeffrey burnette", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7711 + }, + { + "path": "audio-1502195938-headset.flac", + "question": "What begins the list of 'song_name' words in this audio clip?", + "choice_a": "please", + "choice_b": "play", + "choice_c": "thirteen", + "choice_d": "purple haze", + "answer_gt": "thirteen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7712 + }, + { + "path": "audio-1490106229-headset.flac", + "question": "detect the first mention of a word related to 'place_name' in this audio.", + "choice_a": "nigeria", + "choice_b": "the", + "choice_c": "canada", + "choice_d": "china", + "answer_gt": "nigeria", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7713 + }, + { + "path": "audio-1490182522.flac", + "question": "What is the opening word pertaining to 'place_name' mentioned here?", + "choice_a": "london", + "choice_b": "what", + "choice_c": "austin", + "choice_d": "delhi", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7714 + }, + { + "path": "audio-1490799816.flac", + "question": "Which word that signifies 'place_name' is spoken first in this clip?", + "choice_a": "directions", + "choice_b": "to", + "choice_c": "train station", + "choice_d": "disney land", + "answer_gt": "disney land", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7715 + }, + { + "path": "audio-1495734990.flac", + "question": "What's the first word that's in association with 'weather_descriptor' in this recording?", + "choice_a": "i", + "choice_b": "driveway", + "choice_c": "driveway", + "choice_d": "rain or snow", + "answer_gt": "driveway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7716 + }, + { + "path": "audio-1495733739-headset.flac", + "question": "What word relevant to 'transport_type' is mentioned at the start of this audio?", + "choice_a": "tickets", + "choice_b": "for", + "choice_c": "yellow cab", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7717 + }, + { + "path": "audio-1501778771.flac", + "question": "Which term that relates to 'event_name' is spoken first in this clip?", + "choice_a": "officially", + "choice_b": "does", + "choice_c": "super bowl", + "choice_d": "wash the windows", + "answer_gt": "super bowl", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7718 + }, + { + "path": "audio-1501686891-headset.flac", + "question": "detect the first mention of a word related to 'transport_type' in this audio.", + "choice_a": "a", + "choice_b": "ticket", + "choice_c": "cab", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7719 + }, + { + "path": "audio--1504197984-headset.flac", + "question": "What is the initial 'player_setting'-associated word you hear in this recording?", + "choice_a": "to", + "choice_b": "keep", + "choice_c": "keep reading", + "choice_d": "continue reading", + "answer_gt": "keep reading", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7720 + }, + { + "path": "audio-1497267973.flac", + "question": "In this sound clip, which word related to 'business_name' is mentioned first?", + "choice_a": "is", + "choice_b": "googs", + "choice_c": "company name", + "choice_d": "mr pizza", + "answer_gt": "googs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7721 + }, + { + "path": "audio-1490705726-headset.flac", + "question": "What's the first word that's tied to 'music_genre' mentioned in this audio?", + "choice_a": "my", + "choice_b": "rock", + "choice_c": "classic rock", + "choice_d": "pop", + "answer_gt": "rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7722 + }, + { + "path": "audio-1502308809-headset.flac", + "question": "What is the first word related to 'date' mentioned in this audio clip?", + "choice_a": "this", + "choice_b": "going", + "choice_c": "this weekend", + "choice_d": "the twenty third of this month", + "answer_gt": "this weekend", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7723 + }, + { + "path": "audio-1490356135-headset.flac", + "question": "Which word is the first to be related to 'time' in this recording?", + "choice_a": "is", + "choice_b": "when", + "choice_c": "five pm", + "choice_d": "next weeks", + "answer_gt": "five pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7724 + }, + { + "path": "audio-1490957470.flac", + "question": "Which word is the earliest mention of something related to 'place_name' in this audio?", + "choice_a": "are", + "choice_b": "the equator", + "choice_c": "bolshoi theatre", + "choice_d": "bay area", + "answer_gt": "the equator", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7725 + }, + { + "path": "audio-1490704779-headset.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "busy", + "choice_b": "am", + "choice_c": "hemingway", + "choice_d": "charlotte", + "answer_gt": "charlotte", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7726 + }, + { + "path": "audio-1488559509.flac", + "question": "In this audio, what's the first word you hear that's connected to 'personal_info'?", + "choice_a": "add", + "choice_b": "this", + "choice_c": "phone numbers", + "choice_d": "email", + "answer_gt": "email", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7727 + }, + { + "path": "audio-1501686652-headset.flac", + "question": "What is the first word that's associated with 'business_name' in this audio?", + "choice_a": "what", + "choice_b": "terranova", + "choice_c": "company name", + "choice_d": "aircel", + "answer_gt": "terranova", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7728 + }, + { + "path": "audio-1501689922-headset.flac", + "question": "Which word that's evocative of 'date' is mentioned first in this audio?", + "choice_a": "today", + "choice_b": "bad", + "choice_c": "this year", + "choice_d": "wednesday", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7729 + }, + { + "path": "audio-1497871913-headset.flac", + "question": "Which word that's linked to 'event_name' do you hear first in this recording?", + "choice_a": "do", + "choice_b": "appointments", + "choice_c": "haircut", + "choice_d": "appointment", + "answer_gt": "appointments", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7730 + }, + { + "path": "audio-1502374908.flac", + "question": "point out the first 'date'-related word uttered in this audio.", + "choice_a": "the nineteenth", + "choice_b": "what", + "choice_c": "tuesday", + "choice_d": "tomorrow", + "answer_gt": "the nineteenth", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7731 + }, + { + "path": "audio-1495378249.flac", + "question": "What's the first word that's in association with 'transport_type' in this recording?", + "choice_a": "city", + "choice_b": "call", + "choice_c": "yellow cab", + "choice_d": "cab", + "answer_gt": "cab", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7732 + }, + { + "path": "audio-1501754179-headset.flac", + "question": "What's the first term that pertains to 'game_name' in this recording?", + "choice_a": "you", + "choice_b": "parcheese game", + "choice_c": "chess", + "choice_d": "clash of clans", + "answer_gt": "parcheese game", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7733 + }, + { + "path": "audio-1497887741-headset.flac", + "question": "What's the first word that's associated with 'playlist_name' in this clip?", + "choice_a": "wacky", + "choice_b": "me", + "choice_c": "rock list", + "choice_d": "running", + "answer_gt": "wacky", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7734 + }, + { + "path": "audio-1501151452.flac", + "question": "Which word that's evocative of 'music_descriptor' is mentioned first in this audio?", + "choice_a": "play", + "choice_b": "previous", + "choice_c": "christian", + "choice_d": "good song", + "answer_gt": "previous", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7735 + }, + { + "path": "audio-1490200796.flac", + "question": "spot the first word that hints at 'time' in this audio.", + "choice_a": "my", + "choice_b": "seven am", + "choice_c": "thirty minutes", + "choice_d": "five pm", + "answer_gt": "seven am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7736 + }, + { + "path": "audio-1501768428-headset.flac", + "question": "What word is the first to reference 'media_type' in this sound clip?", + "choice_a": "play", + "choice_b": "youtube", + "choice_c": "bbc world", + "choice_d": "so clean", + "answer_gt": "youtube", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7737 + }, + { + "path": "audio-1434538537.flac", + "question": "What's the first term that pertains to 'date' in this recording?", + "choice_a": "is", + "choice_b": "it", + "choice_c": "Friday", + "choice_d": "today's", + "answer_gt": "Friday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7738 + }, + { + "path": "audio-1490368753.flac", + "question": "Which word that signifies 'transport_type' is spoken first in this clip?", + "choice_a": "the", + "choice_b": "to", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7739 + }, + { + "path": "audio-1490019941-headset.flac", + "question": "Which word is introduced first in this audio that relates to 'place_name'?", + "choice_a": "prime", + "choice_b": "was", + "choice_c": "britain", + "choice_d": "alaska", + "answer_gt": "britain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7740 + }, + { + "path": "audio-1501767264.flac", + "question": "find the first word that's linked to 'movie_type' in this sound clip?", + "choice_a": "action", + "choice_b": "do", + "choice_c": "you", + "choice_d": "any", + "answer_gt": "action", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7741 + }, + { + "path": "audio-1497860927.flac", + "question": "Which word is the earliest mention of something related to 'list_name' in this audio?", + "choice_a": "my", + "choice_b": "on", + "choice_c": "grocery", + "choice_d": "todo", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7742 + }, + { + "path": "audio-1434532097-headset.flac", + "question": "What's the first word that's indicative of 'music_genre' in this audio?", + "choice_a": "play", + "choice_b": "techno", + "choice_c": "sad songs", + "choice_d": "smooth jazz", + "answer_gt": "techno", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7743 + }, + { + "path": "audio-1501752132.flac", + "question": "What is the opening word pertaining to 'place_name' mentioned here?", + "choice_a": "world", + "choice_b": "latest", + "choice_c": "las vegas", + "choice_d": "city", + "answer_gt": "world", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7744 + }, + { + "path": "audio-1490289399-headset.flac", + "question": "name the first word that's related to 'player_setting' in this audio?", + "choice_a": "playback", + "choice_b": "clancy", + "choice_c": "continue reading", + "choice_d": "skip", + "answer_gt": "playback", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7745 + }, + { + "path": "audio-1501751649.flac", + "question": "In this sound clip, what's the first word that's relevant to 'person'?", + "choice_a": "old", + "choice_b": "how", + "choice_c": "j. k. rowling", + "choice_d": "george clooney", + "answer_gt": "j. k. rowling", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7746 + }, + { + "path": "audio-1490202363.flac", + "question": "Which word comes first in this audio that's tied to 'event_name'?", + "choice_a": "a", + "choice_b": "i", + "choice_c": "medicine", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7747 + }, + { + "path": "audio-1502374923.flac", + "question": "What's the first term that pertains to 'person' in this recording?", + "choice_a": "please", + "choice_b": "for", + "choice_c": "john smith", + "choice_d": "bruce", + "answer_gt": "bruce", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7748 + }, + { + "path": "audio-1501763783-headset.flac", + "question": "Which word that's synonymous with 'news_topic' appears first in this audio?", + "choice_a": "crime statistics", + "choice_b": "me", + "choice_c": "politics", + "choice_d": "cars", + "answer_gt": "crime statistics", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7749 + }, + { + "path": "audio-1501772656.flac", + "question": "What's the first word that's reflective of 'house_place' in this sound clip?", + "choice_a": "the", + "choice_b": "bathroom", + "choice_c": "hallway", + "choice_d": "living room", + "answer_gt": "bathroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7750 + }, + { + "path": "audio-1497884648.flac", + "question": "Which word comes first in this audio that's tied to 'time'?", + "choice_a": "at", + "choice_b": "am", + "choice_c": "five am", + "choice_d": "five pm to six pm", + "answer_gt": "five am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7751 + }, + { + "path": "audio-1497030905.flac", + "question": "What word is the first to reference 'event_name' in this sound clip?", + "choice_a": "eric's bday", + "choice_b": "delete", + "choice_c": "gun shows", + "choice_d": "conference call", + "answer_gt": "eric's bday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7752 + }, + { + "path": "audio-1499693597.flac", + "question": "find the first word that's linked to 'media_type' in this sound clip?", + "choice_a": "latest", + "choice_b": "twitter", + "choice_c": "audiobook", + "choice_d": "so clean", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7753 + }, + { + "path": "audio-1502893872.flac", + "question": "tell me the first word that's in relation to 'date' in this audio?", + "choice_a": "the", + "choice_b": "set", + "choice_c": "a day before", + "choice_d": "tomorrow", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7754 + }, + { + "path": "audio-1497886177-headset.flac", + "question": "What's the first term that pertains to 'news_topic' in this recording?", + "choice_a": "weather", + "choice_b": "me", + "choice_c": "environmental", + "choice_d": "international women's day", + "answer_gt": "weather", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7755 + }, + { + "path": "audio-1489504031.flac", + "question": "What is the first word related to 'news_topic' mentioned in this audio clip?", + "choice_a": "some", + "choice_b": "business", + "choice_c": "president trump", + "choice_d": "investigation into trump's ties", + "answer_gt": "business", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7756 + }, + { + "path": "audio-1502892952.flac", + "question": "What is the first word that's associated with 'date' in this audio?", + "choice_a": "is", + "choice_b": "the", + "choice_c": "this week", + "choice_d": "monday the twentieth", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7757 + }, + { + "path": "audio--1506078335.flac", + "question": "Which term that relates to 'list_name' is spoken first in this clip?", + "choice_a": "to", + "choice_b": "check", + "choice_c": "stationary", + "choice_d": "to do", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7758 + }, + { + "path": "audio-1501687491.flac", + "question": "Which word is introduced first in this audio that relates to 'currency_name'?", + "choice_a": "is", + "choice_b": "how", + "choice_c": "us dollar", + "choice_d": "british pound", + "answer_gt": "us dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7759 + }, + { + "path": "audio-1502201002.flac", + "question": "What's the first word that's in association with 'date' in this recording?", + "choice_a": "today's", + "choice_b": "forecast", + "choice_c": "Thursday", + "choice_d": "second march", + "answer_gt": "today's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7760 + }, + { + "path": "audio-1490704779.flac", + "question": "What's the first word that's alluding to 'person' in this sound recording?", + "choice_a": "i", + "choice_b": "charlotte", + "choice_c": "jeff", + "choice_d": "j. k. rowling", + "answer_gt": "charlotte", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7761 + }, + { + "path": "audio-1498565005-headset.flac", + "question": "In this sound clip, which word related to 'game_name' is mentioned first?", + "choice_a": "of", + "choice_b": "clash", + "choice_c": "chess", + "choice_d": "clash of clans", + "answer_gt": "clash of clans", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7762 + }, + { + "path": "audio-1502896200-headset.flac", + "question": "name the first word that's related to 'date' in this audio?", + "choice_a": "about", + "choice_b": "this weekend", + "choice_c": "august fifteenth", + "choice_d": "weekdays", + "answer_gt": "this weekend", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7763 + }, + { + "path": "audio-1501152374-headset.flac", + "question": "name the first word that's related to 'list_name' in this audio?", + "choice_a": "grocery", + "choice_b": "please", + "choice_c": "walmart list", + "choice_d": "expenses", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7764 + }, + { + "path": "audio-1490367215-headset.flac", + "question": "What is the opening word pertaining to 'audiobook_name' mentioned here?", + "choice_a": "be more chill", + "choice_b": "to", + "choice_c": "the notebook", + "choice_d": "harry potter", + "answer_gt": "be more chill", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7765 + }, + { + "path": "audio-1497866816-headset.flac", + "question": "What's the first word that's in reference to 'color_type' in this sound clip?", + "choice_a": "pink", + "choice_b": "we", + "choice_c": "blue", + "choice_d": "red", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7766 + }, + { + "path": "audio-1494421815.flac", + "question": "What's the first word that's characteristic of 'radio_name' in this clip?", + "choice_a": "hot one hundred and five", + "choice_b": "and", + "choice_c": "nine hundred and ninety nine fm", + "choice_d": "eight hundred and ninety seven fm", + "answer_gt": "hot one hundred and five", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7767 + }, + { + "path": "audio--1504192750-headset.flac", + "question": "Which word that's synonymous with 'date' appears first in this audio?", + "choice_a": "set", + "choice_b": "today", + "choice_c": "march eighteen", + "choice_d": "next week", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7768 + }, + { + "path": "audio-1498564971-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'event_name'?", + "choice_a": "o'clock", + "choice_b": "meeting", + "choice_c": "medicine", + "choice_d": "call mom", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7769 + }, + { + "path": "audio-1497879959-headset.flac", + "question": "What begins the list of 'person' words in this audio clip?", + "choice_a": "smith", + "choice_b": "find", + "choice_c": "john smith", + "choice_d": "jessica's", + "answer_gt": "john smith", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7770 + }, + { + "path": "audio-1494421815-headset.flac", + "question": "Which word comes first in this audio that's tied to 'radio_name'?", + "choice_a": "hot one hundred and five", + "choice_b": "the", + "choice_c": "station that plays r. n. b.", + "choice_d": "bbc", + "answer_gt": "hot one hundred and five", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7771 + }, + { + "path": "audio-1501774167.flac", + "question": "What's the first word that's in association with 'event_name' in this recording?", + "choice_a": "pay", + "choice_b": "please", + "choice_c": "super bowl", + "choice_d": "pay days", + "answer_gt": "pay days", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7772 + }, + { + "path": "audio-1502890225.flac", + "question": "Which word that's synonymous with 'device_type' appears first in this audio?", + "choice_a": "vacuum", + "choice_b": "the", + "choice_c": "vacuuming", + "choice_d": "all the lights", + "answer_gt": "vacuum", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7773 + }, + { + "path": "audio-1502377426.flac", + "question": "What's the premier word about 'song_name' in this sound clip?", + "choice_a": "lady", + "choice_b": "poker face", + "choice_c": "bohemian raphsody", + "choice_d": "every time i see you", + "answer_gt": "poker face", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7774 + }, + { + "path": "audio-1501756508-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'person'?", + "choice_a": "pink", + "choice_b": "a", + "choice_c": "ajay devgn", + "choice_d": "allearsnet", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7775 + }, + { + "path": "audio--1504198714.flac", + "question": "Which word is the earliest mention of something related to 'time' in this audio?", + "choice_a": "cancel", + "choice_b": "seven am", + "choice_c": "half an hour", + "choice_d": "week", + "answer_gt": "seven am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7776 + }, + { + "path": "audio-1434529788-headset.flac", + "question": "detect the first mention of a word related to 'person' in this audio.", + "choice_a": "contact", + "choice_b": "john's", + "choice_c": "joanna", + "choice_d": "james and alice", + "answer_gt": "john's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7777 + }, + { + "path": "audio-1500978883-headset.flac", + "question": "What begins the list of 'media_type' words in this audio clip?", + "choice_a": "that", + "choice_b": "waiting", + "choice_c": "bbc poll", + "choice_d": "tweet", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7778 + }, + { + "path": "audio-1499265283-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'place_name'?", + "choice_a": "canada", + "choice_b": "current", + "choice_c": "german", + "choice_d": "city", + "answer_gt": "canada", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7779 + }, + { + "path": "audio-1495378287-headset.flac", + "question": "What's the premier word about 'joke_type' in this sound clip?", + "choice_a": "what's", + "choice_b": "joke", + "choice_c": "funniest", + "choice_d": "funny", + "answer_gt": "funniest", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7780 + }, + { + "path": "audio-1499692723.flac", + "question": "Which word that's linked to 'transport_type' do you hear first in this recording?", + "choice_a": "call", + "choice_b": "call", + "choice_c": "cab", + "choice_d": "taxi", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7781 + }, + { + "path": "audio-1501752435.flac", + "question": "In this audio clip, what's the first word that pertains to 'game_name'?", + "choice_a": "clash of clans", + "choice_b": "begin", + "choice_c": "parcheese game", + "choice_d": "battleship", + "answer_gt": "clash of clans", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7782 + }, + { + "path": "audio-1492783720.flac", + "question": "What's the first word that's representative of 'playlist_name' in this recording?", + "choice_a": "the", + "choice_b": "itunes library", + "choice_c": "summer jams playlist", + "choice_d": "rock list", + "answer_gt": "itunes library", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7783 + }, + { + "path": "audio-1490193801-headset.flac", + "question": "What's the first word that's alluding to 'player_setting' in this sound recording?", + "choice_a": "jump", + "choice_b": "to", + "choice_c": "stop play", + "choice_d": "playback", + "answer_gt": "jump", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7784 + }, + { + "path": "audio-1502194762.flac", + "question": "What's the earliest 'weather_descriptor'-related term you hear in this clip?", + "choice_a": "rain", + "choice_b": "the", + "choice_c": "rain or snow", + "choice_d": "sunscreen", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7785 + }, + { + "path": "audio-1501765013.flac", + "question": "What's the first word that comes up in this audio in relation to 'place_name'?", + "choice_a": "in", + "choice_b": "beijing", + "choice_c": "bay area", + "choice_d": "earth", + "answer_gt": "beijing", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7786 + }, + { + "path": "audio-1501407272.flac", + "question": "What is the initial 'media_type'-associated word you hear in this recording?", + "choice_a": "the", + "choice_b": "twitter", + "choice_c": "social networks", + "choice_d": "so clean", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7787 + }, + { + "path": "audio-1492784618.flac", + "question": "What's the earliest 'artist_name'-related term you hear in this clip?", + "choice_a": "the", + "choice_b": "me", + "choice_c": "adele", + "choice_d": "the beatles", + "answer_gt": "the beatles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7788 + }, + { + "path": "audio-1498758038.flac", + "question": "name the first word that's related to 'definition_word' in this audio?", + "choice_a": "sphere", + "choice_b": "a", + "choice_c": "smartphone circuit", + "choice_d": "motivation", + "answer_gt": "sphere", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7789 + }, + { + "path": "audio-1490019770.flac", + "question": "What is the initial 'media_type'-associated word you hear in this recording?", + "choice_a": "that", + "choice_b": "podcasts", + "choice_c": "new york time's", + "choice_d": "time of india", + "answer_gt": "podcasts", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7790 + }, + { + "path": "audio--1506077996.flac", + "question": "spot the first word that hints at 'music_genre' in this audio.", + "choice_a": "want", + "choice_b": "i", + "choice_c": "smooth jazz", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7791 + }, + { + "path": "audio--1505405792.flac", + "question": "In this sound clip, what's the first word that's relevant to 'player_setting'?", + "choice_a": "to", + "choice_b": "audiobook", + "choice_c": "keep reading", + "choice_d": "put on", + "answer_gt": "keep reading", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7792 + }, + { + "path": "audio-1494418675.flac", + "question": "What is the first word that's associated with 'event_name' in this audio?", + "choice_a": "reminder", + "choice_b": "notification", + "choice_c": "meeting", + "choice_d": "second of next mont lands", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7793 + }, + { + "path": "audio-1490368787.flac", + "question": "In this sound clip, what's the first word that's relevant to 'date'?", + "choice_a": "today", + "choice_b": "olly", + "choice_c": "this week's", + "choice_d": "tomorrows", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7794 + }, + { + "path": "audio-1498484889.flac", + "question": "What's the first word that's alluding to 'artist_name' in this sound recording?", + "choice_a": "third", + "choice_b": "music", + "choice_c": "mf doom", + "choice_d": "third day", + "answer_gt": "third day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7795 + }, + { + "path": "audio-1497868704.flac", + "question": "What word relevant to 'place_name' is mentioned at the start of this audio?", + "choice_a": "time", + "choice_b": "the", + "choice_c": "canada", + "choice_d": "australia", + "answer_gt": "australia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7796 + }, + { + "path": "audio-1500979277-headset.flac", + "question": "What's the first word that's associated with 'place_name' in this clip?", + "choice_a": "town", + "choice_b": "around", + "choice_c": "earth", + "choice_d": "chennai", + "answer_gt": "town", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7797 + }, + { + "path": "audio-1490706309.flac", + "question": "point out the first 'list_name'-related word uttered in this audio.", + "choice_a": "repeat", + "choice_b": "shopping", + "choice_c": "todo", + "choice_d": "grocery", + "answer_gt": "shopping", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7798 + }, + { + "path": "audio-1497860774.flac", + "question": "Which word linked to 'definition_word' comes first in this recording?", + "choice_a": "sphere", + "choice_b": "describe", + "choice_c": "converse shoes", + "choice_d": "elaborate", + "answer_gt": "sphere", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7799 + }, + { + "path": "audio-1488973561.flac", + "question": "Which word is introduced first in this audio that relates to 'personal_info'?", + "choice_a": "add", + "choice_b": "person's", + "choice_c": "phone numbers", + "choice_d": "email", + "answer_gt": "email", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7800 + }, + { + "path": "audio-1502896000.flac", + "question": "What begins the list of 'food_type' words in this audio clip?", + "choice_a": "pack", + "choice_b": "noodles", + "choice_c": "tuna", + "choice_d": "sushi", + "answer_gt": "noodles", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7801 + }, + { + "path": "audio--1504190955.flac", + "question": "What is the opening word pertaining to 'player_setting' mentioned here?", + "choice_a": "resume", + "choice_b": "animal", + "choice_c": "read", + "choice_d": "skip", + "answer_gt": "resume", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7802 + }, + { + "path": "audio-1498570470-headset.flac", + "question": "tell me the first word that's in relation to 'person' in this audio?", + "choice_a": "did", + "choice_b": "elvis", + "choice_c": "john doe", + "choice_d": "dan", + "answer_gt": "elvis", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7803 + }, + { + "path": "audio-1498569267.flac", + "question": "Which word that's evocative of 'currency_name' is mentioned first in this audio?", + "choice_a": "compared", + "choice_b": "worth", + "choice_c": "british pound", + "choice_d": "dollar", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7804 + }, + { + "path": "audio-1497439388-headset.flac", + "question": "Which word linked to 'device_type' comes first in this recording?", + "choice_a": "wemo", + "choice_b": "power", + "choice_c": "coffee pot", + "choice_d": "all the lights", + "answer_gt": "wemo", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7805 + }, + { + "path": "audio-1501686161-headset.flac", + "question": "In this sound clip, which word related to 'event_name' is mentioned first?", + "choice_a": "a", + "choice_b": "with", + "choice_c": "meeting", + "choice_d": "eric's bday", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7806 + }, + { + "path": "audio-1501758745.flac", + "question": "What is the first word related to 'date' mentioned in this audio clip?", + "choice_a": "are", + "choice_b": "todays", + "choice_c": "two days", + "choice_d": "this weekend", + "answer_gt": "todays", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7807 + }, + { + "path": "audio-1499694147-headset.flac", + "question": "What's the first word that's tied to 'music_genre' mentioned in this audio?", + "choice_a": "play", + "choice_b": "nineties rap", + "choice_c": "classic rock", + "choice_d": "techno", + "answer_gt": "nineties rap", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7808 + }, + { + "path": "audio-1498484889-headset.flac", + "question": "What's the earliest 'artist_name'-related term you hear in this clip?", + "choice_a": "third", + "choice_b": "third day", + "choice_c": "adele", + "choice_d": "david bowie", + "answer_gt": "third day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7809 + }, + { + "path": "audio-1499693115.flac", + "question": "Which word comes first in this audio that's tied to 'weather_descriptor'?", + "choice_a": "jacket", + "choice_b": "a", + "choice_c": "rain", + "choice_d": "warm socks", + "answer_gt": "jacket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7810 + }, + { + "path": "audio-1501697044-headset.flac", + "question": "Which word comes first in this audio that's tied to 'time'?", + "choice_a": "nine am", + "choice_b": "me", + "choice_c": "week", + "choice_d": "seven am", + "answer_gt": "nine am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7811 + }, + { + "path": "audio-1502218507.flac", + "question": "Which word that's synonymous with 'date' appears first in this audio?", + "choice_a": "thirteenth june", + "choice_b": "set", + "choice_c": "this week's", + "choice_d": "next month", + "answer_gt": "thirteenth june", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7812 + }, + { + "path": "audio-1488552005.flac", + "question": "In this sound clip, what's the first word that's relevant to 'event_name'?", + "choice_a": "from", + "choice_b": "calendar", + "choice_c": "dentist appointment", + "choice_d": "meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7813 + }, + { + "path": "audio-1502896988.flac", + "question": "What's the premier word about 'house_place' in this sound clip?", + "choice_a": "room", + "choice_b": "off", + "choice_c": "house", + "choice_d": "bedroom", + "answer_gt": "room", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7814 + }, + { + "path": "audio-1488538285.flac", + "question": "What's the first word that's indicative of 'place_name' in this audio?", + "choice_a": "international", + "choice_b": "me", + "choice_c": "home", + "choice_d": "international", + "answer_gt": "international", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7815 + }, + { + "path": "audio-1494547839.flac", + "question": "Which term that relates to 'food_type' is spoken first in this clip?", + "choice_a": "order", + "choice_b": "for", + "choice_c": "sushi", + "choice_d": "pav bhaji", + "answer_gt": "sushi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7816 + }, + { + "path": "audio-1501752485-headset.flac", + "question": "Which word is the first to be related to 'event_name' in this recording?", + "choice_a": "meeting", + "choice_b": "schedule", + "choice_c": "to have tablets", + "choice_d": "fringe", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7817 + }, + { + "path": "audio-1501752104.flac", + "question": "What's the first word mentioned in this audio that's about 'house_place'?", + "choice_a": "the", + "choice_b": "turn", + "choice_c": "kitchen", + "choice_d": "house", + "answer_gt": "kitchen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7818 + }, + { + "path": "audio-1490007224-headset.flac", + "question": "What is the first word that's associated with 'place_name' in this audio?", + "choice_a": "the", + "choice_b": "london", + "choice_c": "miami", + "choice_d": "earth", + "answer_gt": "london", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7819 + }, + { + "path": "audio-1498572042.flac", + "question": "Which word is introduced first in this audio that relates to 'date'?", + "choice_a": "for", + "choice_b": "date", + "choice_c": "tomorrows", + "choice_d": "tuesday", + "answer_gt": "tuesday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7820 + }, + { + "path": "audio-1488558888.flac", + "question": "What's the earliest 'list_name'-related term you hear in this clip?", + "choice_a": "what's", + "choice_b": "on", + "choice_c": "todo", + "choice_d": "party time", + "answer_gt": "todo", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7821 + }, + { + "path": "audio-1502376123-headset.flac", + "question": "Which word is the earliest mention of something related to 'list_name' in this audio?", + "choice_a": "do", + "choice_b": "new", + "choice_c": "to do", + "choice_d": "favorite albums", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7822 + }, + { + "path": "audio-1502099804-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'date'?", + "choice_a": "to", + "choice_b": "fifth march", + "choice_c": "todays", + "choice_d": "yesterday", + "answer_gt": "fifth march", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7823 + }, + { + "path": "audio-1490007947-headset.flac", + "question": "What is the first word that's associated with 'place_name' in this audio?", + "choice_a": "this", + "choice_b": "downtown", + "choice_c": "my place", + "choice_d": "jaipur", + "answer_gt": "downtown", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7824 + }, + { + "path": "audio-1490967291-headset.flac", + "question": "tell me the first 'event_name'-connected word in this audio.", + "choice_a": "documents", + "choice_b": "print the documents", + "choice_c": "pick up my parents", + "choice_d": "appointments", + "answer_gt": "print the documents", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7825 + }, + { + "path": "audio-1494607714-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'house_place'?", + "choice_a": "house", + "choice_b": "you", + "choice_c": "house", + "choice_d": "kitchen", + "answer_gt": "house", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7826 + }, + { + "path": "audio-1497861026-headset.flac", + "question": "tell me the first 'person'-connected word in this audio.", + "choice_a": "mary's", + "choice_b": "is", + "choice_c": "steph curry", + "choice_d": "j. k. rowling", + "answer_gt": "mary's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7827 + }, + { + "path": "audio-1501690324.flac", + "question": "What's the first word mentioned in this audio that's about 'definition_word'?", + "choice_a": "description", + "choice_b": "find", + "choice_c": "happy birthday", + "choice_d": "samsung tv", + "answer_gt": "samsung tv", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7828 + }, + { + "path": "audio-1488992997.flac", + "question": "What is the first word that's in connection with 'time' in this audio?", + "choice_a": "new", + "choice_b": "in", + "choice_c": "last ten minutes", + "choice_d": "thirty minute", + "answer_gt": "last ten minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7829 + }, + { + "path": "audio-1488530906.flac", + "question": "Which word is the earliest mention of something related to 'event_name' in this audio?", + "choice_a": "you", + "choice_b": "what", + "choice_c": "appointments", + "choice_d": "fringe", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7830 + }, + { + "path": "audio-1499361558.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "to", + "choice_b": "sent", + "choice_c": "jane", + "choice_d": "naty", + "answer_gt": "jane", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7831 + }, + { + "path": "audio-1498570797-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'media_type'?", + "choice_a": "i'm", + "choice_b": "post", + "choice_c": "facebook", + "choice_d": "bbc poll", + "answer_gt": "facebook", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7832 + }, + { + "path": "audio-1501772231-headset.flac", + "question": "What is the first word that's in connection with 'date' in this audio?", + "choice_a": "this week", + "choice_b": "me", + "choice_c": "second march", + "choice_d": "next week", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7833 + }, + { + "path": "audio-1497439036.flac", + "question": "In this audio, what's the first word you hear that's connected to 'media_type'?", + "choice_a": "the", + "choice_b": "cbs", + "choice_c": "twitter", + "choice_d": "audiobook", + "answer_gt": "cbs", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7834 + }, + { + "path": "audio-1498568298-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'transport_type'?", + "choice_a": "uberpool", + "choice_b": "island", + "choice_c": "yellow cab", + "choice_d": "metro", + "answer_gt": "uberpool", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7835 + }, + { + "path": "audio-1499693974-headset.flac", + "question": "Which term that relates to 'artist_name' is spoken first in this clip?", + "choice_a": "adele's", + "choice_b": "my", + "choice_c": "celine dion", + "choice_d": "ella fitzgerals", + "answer_gt": "adele's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7836 + }, + { + "path": "audio-1501151625.flac", + "question": "What's the first word that comes up in this audio in relation to 'color_type'?", + "choice_a": "to", + "choice_b": "more", + "choice_c": "pink", + "choice_d": "red ish", + "answer_gt": "red ish", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7837 + }, + { + "path": "audio--1505404838.flac", + "question": "Which word that's affiliated with 'radio_name' is spoken first in this clip?", + "choice_a": "station that plays r. n. b.", + "choice_b": "station", + "choice_c": "hot one hundred and five", + "choice_d": "sirius xm radio channel", + "answer_gt": "station that plays r. n. b.", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7838 + }, + { + "path": "audio-1488991681.flac", + "question": "What's the first word that's reflective of 'person' in this sound clip?", + "choice_a": "sent", + "choice_b": "to", + "choice_c": "jerry", + "choice_d": "jane", + "answer_gt": "jane", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7839 + }, + { + "path": "audio-1490195852.flac", + "question": "What's the first word that comes up in this audio in relation to 'timeofday'?", + "choice_a": "can", + "choice_b": "tonight", + "choice_c": "noon", + "choice_d": "this afternoon", + "answer_gt": "tonight", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7840 + }, + { + "path": "audio-1502377642-headset.flac", + "question": "What's the first word that's connected to 'podcast_descriptor' in this audio?", + "choice_a": "listening to yesterday", + "choice_b": "i", + "choice_c": "next podcast episode", + "choice_d": "started last night", + "answer_gt": "listening to yesterday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7841 + }, + { + "path": "audio-1490359729-headset.flac", + "question": "What word is the first to reference 'coffee_type' in this sound clip?", + "choice_a": "without", + "choice_b": "please", + "choice_c": "creamy", + "choice_d": "without sweetener", + "answer_gt": "without sweetener", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7842 + }, + { + "path": "audio-1500388626-headset.flac", + "question": "tell me the first 'transport_type'-connected word in this audio.", + "choice_a": "go", + "choice_b": "how", + "choice_c": "metro", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7843 + }, + { + "path": "audio-1497264963-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'person'?", + "choice_a": "jerry", + "choice_b": "time", + "choice_c": "patrick stewart", + "choice_d": "jane's", + "answer_gt": "jerry", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7844 + }, + { + "path": "audio-1501755987.flac", + "question": "What is the first word that's in connection with 'music_genre' in this audio?", + "choice_a": "find", + "choice_b": "smooth jazz", + "choice_c": "pop", + "choice_d": "techno", + "answer_gt": "smooth jazz", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7845 + }, + { + "path": "audio-1490198407-headset.flac", + "question": "What's the first word that's alluding to 'media_type' in this sound recording?", + "choice_a": "new", + "choice_b": "anything", + "choice_c": "insta", + "choice_d": "new york time's", + "answer_gt": "insta", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7846 + }, + { + "path": "audio-1498568666.flac", + "question": "detect the first mention of a word related to 'person' in this audio.", + "choice_a": "will", + "choice_b": "kate's", + "choice_c": "trump", + "choice_d": "john", + "answer_gt": "kate's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7847 + }, + { + "path": "audio-1502301982.flac", + "question": "What's the first word mentioned in this audio that's about 'event_name'?", + "choice_a": "meetings", + "choice_b": "i", + "choice_c": "parade", + "choice_d": "easter", + "answer_gt": "meetings", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7848 + }, + { + "path": "audio-1488984518.flac", + "question": "What's the first word in this audio that has a connection to 'person'?", + "choice_a": "joanna", + "choice_b": "meeting", + "choice_c": "john doe", + "choice_d": "morgan's", + "answer_gt": "joanna", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7849 + }, + { + "path": "audio-1498572820.flac", + "question": "What is the first word in this clip that's relevant to 'date'?", + "choice_a": "at", + "choice_b": "tomorrow", + "choice_c": "next week", + "choice_d": "this week", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7850 + }, + { + "path": "audio-1502376743.flac", + "question": "Which word that's linked to 'currency_name' do you hear first in this recording?", + "choice_a": "pounds", + "choice_b": "are", + "choice_c": "dollar", + "choice_d": "u. s. d.", + "answer_gt": "pounds", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7851 + }, + { + "path": "audio-1490357077-headset.flac", + "question": "detect the first mention of a word related to 'house_place' in this audio.", + "choice_a": "the", + "choice_b": "floor", + "choice_c": "patio", + "choice_d": "garage", + "answer_gt": "floor", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7852 + }, + { + "path": "audio-1498564518-headset.flac", + "question": "In this sound clip, which word related to 'definition_word' is mentioned first?", + "choice_a": "hell", + "choice_b": "describe", + "choice_c": "rock sand", + "choice_d": "smartphone circuit", + "answer_gt": "hell", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7853 + }, + { + "path": "audio-1497622747.flac", + "question": "What's the first word that's associated with 'time' in this clip?", + "choice_a": "me", + "choice_b": "half an hour", + "choice_c": "six am", + "choice_d": "fifteen minutes", + "answer_gt": "half an hour", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7854 + }, + { + "path": "audio-1501151963-headset.flac", + "question": "What's the first word that's tied to 'weather_descriptor' mentioned in this audio?", + "choice_a": "it", + "choice_b": "rain", + "choice_c": "driveway", + "choice_d": "nice", + "answer_gt": "rain", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7855 + }, + { + "path": "audio-1495370134-headset.flac", + "question": "point out the first 'event_name'-related word uttered in this audio.", + "choice_a": "for", + "choice_b": "meeting", + "choice_c": "birthday", + "choice_d": "church services", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7856 + }, + { + "path": "audio-1492785589-headset.flac", + "question": "What is the initial 'list_name'-associated word you hear in this recording?", + "choice_a": "check", + "choice_b": "recent", + "choice_c": "activities", + "choice_d": "to do", + "answer_gt": "to do", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7857 + }, + { + "path": "audio-1495376975-headset.flac", + "question": "What is the first word that's in connection with 'place_name' in this audio?", + "choice_a": "to", + "choice_b": "location", + "choice_c": "international", + "choice_d": "las vegas", + "answer_gt": "location", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7858 + }, + { + "path": "audio-1434539121.flac", + "question": "In this sound clip, which word related to 'event_name' is mentioned first?", + "choice_a": "this", + "choice_b": "parade", + "choice_c": "registration", + "choice_d": "medical appointment", + "answer_gt": "parade", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7859 + }, + { + "path": "audio-1502375443-headset.flac", + "question": "In this sound clip, which word related to 'weather_descriptor' is mentioned first?", + "choice_a": "the", + "choice_b": "temperature", + "choice_c": "full cover jacket", + "choice_d": "jacket", + "answer_gt": "temperature", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7860 + }, + { + "path": "audio-1497880279.flac", + "question": "point out the first 'definition_word'-related word uttered in this audio.", + "choice_a": "me", + "choice_b": "description", + "choice_c": "rock sand", + "choice_d": "television circuit", + "answer_gt": "television circuit", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7861 + }, + { + "path": "audio-1492779278-headset.flac", + "question": "What's the first word that's alluding to 'date' in this sound recording?", + "choice_a": "what", + "choice_b": "today", + "choice_c": "today", + "choice_d": "march eighteen", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7862 + }, + { + "path": "audio-1498567301.flac", + "question": "tell me the first word that's in relation to 'place_name' in this audio?", + "choice_a": "the", + "choice_b": "california", + "choice_c": "paris", + "choice_d": "india", + "answer_gt": "california", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7863 + }, + { + "path": "audio-1498571876.flac", + "question": "Which word that's linked to 'date' do you hear first in this recording?", + "choice_a": "have", + "choice_b": "today", + "choice_c": "a day before", + "choice_d": "today's", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7864 + }, + { + "path": "audio-1500978995.flac", + "question": "What's the first word that's indicative of 'place_name' in this audio?", + "choice_a": "russia", + "choice_b": "the", + "choice_c": "library", + "choice_d": "the equator", + "answer_gt": "russia", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7865 + }, + { + "path": "audio-1501764367-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'list_name'?", + "choice_a": "grocery", + "choice_b": "to", + "choice_c": "old english", + "choice_d": "groceries", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7866 + }, + { + "path": "audio-1501772022-headset.flac", + "question": "Which term that relates to 'currency_name' is spoken first in this clip?", + "choice_a": "dollar", + "choice_b": "rate", + "choice_c": "dolla", + "choice_d": "us", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7867 + }, + { + "path": "audio-1497368916.flac", + "question": "name the first word that's related to 'music_descriptor' in this audio?", + "choice_a": "some", + "choice_b": "christian", + "choice_c": "good song", + "choice_d": "sad", + "answer_gt": "christian", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7868 + }, + { + "path": "audio-1501155749-headset.flac", + "question": "Which word that's linked to 'relation' do you hear first in this recording?", + "choice_a": "talk", + "choice_b": "need", + "choice_c": "coworker", + "choice_d": "mom", + "answer_gt": "mom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7869 + }, + { + "path": "audio-1490104008.flac", + "question": "What's the first word that comes up in this audio in relation to 'place_name'?", + "choice_a": "show", + "choice_b": "sacramento", + "choice_c": "town", + "choice_d": "las vegas", + "answer_gt": "sacramento", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7870 + }, + { + "path": "audio-1494606930-headset.flac", + "question": "What is the first word that's in connection with 'place_name' in this audio?", + "choice_a": "new york city", + "choice_b": "is", + "choice_c": "orange tx", + "choice_d": "japan", + "answer_gt": "new york city", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7871 + }, + { + "path": "audio-1501759157-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'business_name'?", + "choice_a": "find", + "choice_b": "steakhouse", + "choice_c": "charley's steakhouse", + "choice_d": "i. b. m.", + "answer_gt": "charley's steakhouse", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7872 + }, + { + "path": "audio-1502309018-headset.flac", + "question": "What's the premier word about 'event_name' in this sound clip?", + "choice_a": "when", + "choice_b": "meeting", + "choice_c": "wedding", + "choice_d": "pay days", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7873 + }, + { + "path": "audio-1494608437.flac", + "question": "Which word linked to 'player_setting' comes first in this recording?", + "choice_a": "resume", + "choice_b": "to", + "choice_c": "continue", + "choice_d": "keep playing", + "answer_gt": "resume", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7874 + }, + { + "path": "audio-1490184678.flac", + "question": "What's the first word mentioned in this audio that's about 'transport_type'?", + "choice_a": "go", + "choice_b": "to", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7875 + }, + { + "path": "audio-1501763753.flac", + "question": "find the first word that's linked to 'media_type' in this sound clip?", + "choice_a": "check", + "choice_b": "social", + "choice_c": "social networks", + "choice_d": "youtube", + "answer_gt": "social networks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7876 + }, + { + "path": "audio-1501407251.flac", + "question": "point out the first 'meal_type'-related word uttered in this audio.", + "choice_a": "ready", + "choice_b": "dinner", + "choice_c": "dinner", + "choice_d": "lunch", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7877 + }, + { + "path": "audio-1490703265-headset.flac", + "question": "What is the first word in this clip that's relevant to 'music_genre'?", + "choice_a": "play", + "choice_b": "my", + "choice_c": "sad songs", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7878 + }, + { + "path": "audio-1490194432.flac", + "question": "What's the first word that's connected to 'list_name' in this audio?", + "choice_a": "grocery", + "choice_b": "milk", + "choice_c": "favorite albums", + "choice_d": "groceries", + "answer_gt": "grocery", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7879 + }, + { + "path": "audio-1502374923-headset.flac", + "question": "What's the first word that's representative of 'person' in this recording?", + "choice_a": "bruce", + "choice_b": "load", + "choice_c": "susan", + "choice_d": "hemingway", + "answer_gt": "bruce", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7880 + }, + { + "path": "audio-1502197525-headset.flac", + "question": "What's the first word that's suggestive of 'media_type' in this audio recording?", + "choice_a": "podcasts", + "choice_b": "men's", + "choice_c": "tweet", + "choice_d": "audio book", + "answer_gt": "podcasts", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7881 + }, + { + "path": "audio-1502309637.flac", + "question": "Which word that's affiliated with 'person' is spoken first in this clip?", + "choice_a": "smith", + "choice_b": "john", + "choice_c": "mike", + "choice_d": "john smith", + "answer_gt": "john smith", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7882 + }, + { + "path": "audio-1498567847.flac", + "question": "What's the first term that pertains to 'meal_type' in this recording?", + "choice_a": "should", + "choice_b": "make", + "choice_c": "lunch", + "choice_d": "dinner", + "answer_gt": "dinner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7883 + }, + { + "path": "audio-1498573617.flac", + "question": "detect the first mention of a word related to 'order_type' in this audio.", + "choice_a": "they", + "choice_b": "can", + "choice_c": "deliver", + "choice_d": "takeaway", + "answer_gt": "takeaway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7884 + }, + { + "path": "audio-1501758842-headset.flac", + "question": "Which word comes first in this audio that's tied to 'weather_descriptor'?", + "choice_a": "me", + "choice_b": "i", + "choice_c": "jacket", + "choice_d": "raincoat", + "answer_gt": "raincoat", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7885 + }, + { + "path": "audio-1501778472-headset.flac", + "question": "What's the earliest 'person'-related term you hear in this clip?", + "choice_a": "email", + "choice_b": "ben's", + "choice_c": "donald trump", + "choice_d": "james and alice", + "answer_gt": "ben's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7886 + }, + { + "path": "audio-1501692440-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'time'?", + "choice_a": "it", + "choice_b": "standard", + "choice_c": "thirty minutes", + "choice_d": "seven am", + "answer_gt": "seven am", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7887 + }, + { + "path": "audio-1490706104-headset.flac", + "question": "Which term that relates to 'date' is spoken first in this clip?", + "choice_a": "today", + "choice_b": "siri", + "choice_c": "the nineteenth", + "choice_d": "the twenty third of this month", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7888 + }, + { + "path": "audio-1502893975-headset.flac", + "question": "What is the first word related to 'house_place' mentioned in this audio clip?", + "choice_a": "the", + "choice_b": "floor", + "choice_c": "kitchen", + "choice_d": "hallway", + "answer_gt": "floor", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7889 + }, + { + "path": "audio-1490106525.flac", + "question": "What is the initial 'event_name'-associated word you hear in this recording?", + "choice_a": "meeting", + "choice_b": "put", + "choice_c": "second of next mont lands", + "choice_d": "lunch meeting", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7890 + }, + { + "path": "audio-1502894645-headset.flac", + "question": "What's the premier word about 'media_type' in this sound clip?", + "choice_a": "@potus", + "choice_b": "twitter", + "choice_c": "youtube", + "choice_d": "twitter", + "answer_gt": "twitter", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7891 + }, + { + "path": "audio-1497555803-headset.flac", + "question": "spot the first word that hints at 'time' in this audio.", + "choice_a": "three thirty pm", + "choice_b": "on", + "choice_c": "eight am", + "choice_d": "ten am", + "answer_gt": "three thirty pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7892 + }, + { + "path": "audio-1498570661-headset.flac", + "question": "What's the first word that comes up in this audio in relation to 'weather_descriptor'?", + "choice_a": "snow tyres", + "choice_b": "change", + "choice_c": "temperature", + "choice_d": "raincoat", + "answer_gt": "snow tyres", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7893 + }, + { + "path": "audio-1495734990-headset.flac", + "question": "What's the first word that's tied to 'weather_descriptor' mentioned in this audio?", + "choice_a": "driveway", + "choice_b": "need", + "choice_c": "snow tyres", + "choice_d": "sweater", + "answer_gt": "driveway", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7894 + }, + { + "path": "audio-1497886758-headset.flac", + "question": "tell me the first word that's in relation to 'media_type' in this audio?", + "choice_a": "bob's news", + "choice_b": "bob's", + "choice_c": "insta", + "choice_d": "cnn", + "answer_gt": "bob's news", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7895 + }, + { + "path": "audio-1500036864.flac", + "question": "point out the first 'playlist_name'-related word uttered in this audio.", + "choice_a": "to", + "choice_b": "running", + "choice_c": "rock list", + "choice_d": "summer jams playlist", + "answer_gt": "running", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7896 + }, + { + "path": "audio-1495371776.flac", + "question": "Which word is introduced first in this audio that relates to 'date'?", + "choice_a": "next week", + "choice_b": "what", + "choice_c": "a day before", + "choice_d": "tuesday", + "answer_gt": "next week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7897 + }, + { + "path": "audio-1498574094-headset.flac", + "question": "Which word that's synonymous with 'place_name' appears first in this audio?", + "choice_a": "international", + "choice_b": "read", + "choice_c": "la", + "choice_d": "new york", + "answer_gt": "international", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7898 + }, + { + "path": "audio-1502200984.flac", + "question": "What is the first word that's in connection with 'media_type' in this audio?", + "choice_a": "guardian", + "choice_b": "from", + "choice_c": "tweet", + "choice_d": "podcasts", + "answer_gt": "guardian", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7899 + }, + { + "path": "audio-1501689346.flac", + "question": "name the first word that's related to 'device_type' in this audio?", + "choice_a": "vacuum", + "choice_b": "vacuum cleaner", + "choice_c": "lamps", + "choice_d": "plug", + "answer_gt": "vacuum cleaner", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7900 + }, + { + "path": "audio-1498484053-headset.flac", + "question": "What word is the first to reference 'person' in this sound clip?", + "choice_a": "new", + "choice_b": "emails", + "choice_c": "sarah", + "choice_d": "julie smith", + "answer_gt": "sarah", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7901 + }, + { + "path": "audio-1500037209.flac", + "question": "What's the first word in this audio that has a connection to 'place_name'?", + "choice_a": "tell", + "choice_b": "about", + "choice_c": "las vegas", + "choice_d": "india's", + "answer_gt": "india's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7902 + }, + { + "path": "audio-1490019414.flac", + "question": "Which word comes first in this audio that's tied to 'date'?", + "choice_a": "today", + "choice_b": "date", + "choice_c": "next month", + "choice_d": "tomorrow's", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7903 + }, + { + "path": "audio-1498484329-headset.flac", + "question": "What's the first word that's reflective of 'event_name' in this sound clip?", + "choice_a": "meeting", + "choice_b": "set", + "choice_c": "wedding", + "choice_d": "conference call", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7904 + }, + { + "path": "audio-1501773443.flac", + "question": "Which word that's affiliated with 'media_type' is spoken first in this clip?", + "choice_a": "news", + "choice_b": "cnn", + "choice_c": "new york time's", + "choice_d": "tweet", + "answer_gt": "cnn", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7905 + }, + { + "path": "audio-1488973672.flac", + "question": "What's the first word that's in reference to 'list_name' in this sound clip?", + "choice_a": "start", + "choice_b": "my", + "choice_c": "kitchen", + "choice_d": "walmart list", + "answer_gt": "kitchen", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7906 + }, + { + "path": "audio-1501756059.flac", + "question": "What's the first word that's connected to 'event_name' in this audio?", + "choice_a": "payment", + "choice_b": "car payment", + "choice_c": "dentist appointment", + "choice_d": "conference call", + "answer_gt": "car payment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7907 + }, + { + "path": "audio-1494418533-headset.flac", + "question": "What's the first word that's indicative of 'music_genre' in this audio?", + "choice_a": "jazz", + "choice_b": "like", + "choice_c": "nineties rap", + "choice_d": "rap", + "answer_gt": "jazz", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7908 + }, + { + "path": "audio-1501778688.flac", + "question": "What's the first word that's indicative of 'list_name' in this audio?", + "choice_a": "old", + "choice_b": "old english", + "choice_c": "stationary", + "choice_d": "activities", + "answer_gt": "old english", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7909 + }, + { + "path": "audio-1502894879.flac", + "question": "What's the first word mentioned in this audio that's about 'person'?", + "choice_a": "an", + "choice_b": "to", + "choice_c": "hayathis", + "choice_d": "chelsea", + "answer_gt": "chelsea", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7910 + }, + { + "path": "audio-1501751808-headset.flac", + "question": "Which term that relates to 'event_name' is spoken first in this clip?", + "choice_a": "remove", + "choice_b": "the", + "choice_c": "medicine", + "choice_d": "appointment", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7911 + }, + { + "path": "audio-1497024858-headset.flac", + "question": "Which word that's synonymous with 'cooking_type' appears first in this audio?", + "choice_a": "should", + "choice_b": "the", + "choice_c": "boil", + "choice_d": "boiling", + "answer_gt": "boil", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7912 + }, + { + "path": "audio-1501752730.flac", + "question": "What word is the first to reference 'food_type' in this sound clip?", + "choice_a": "find", + "choice_b": "chocolate", + "choice_c": "chocolate cake", + "choice_d": "red velvet cake", + "answer_gt": "chocolate cake", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7913 + }, + { + "path": "audio-1502375284-headset.flac", + "question": "What is the initial 'event_name'-associated word you hear in this recording?", + "choice_a": "parks", + "choice_b": "some", + "choice_c": "theme parks", + "choice_d": "pay days", + "answer_gt": "theme parks", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7914 + }, + { + "path": "audio-1490957695-headset.flac", + "question": "identify the initial word associated with 'media_type' in this audio.", + "choice_a": "nexus", + "choice_b": "new", + "choice_c": "podcasts", + "choice_d": "tweet", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7915 + }, + { + "path": "audio-1434538853.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "petey's", + "choice_b": "give", + "choice_c": "selena gomez", + "choice_d": "bob", + "answer_gt": "petey's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7916 + }, + { + "path": "audio-1501764461.flac", + "question": "In this audio, what's the first word you hear that's connected to 'place_name'?", + "choice_a": "the", + "choice_b": "and", + "choice_c": "my place", + "choice_d": "one hundred and one", + "answer_gt": "one hundred and one", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7917 + }, + { + "path": "audio-1498567418.flac", + "question": "Which word that's synonymous with 'player_setting' appears first in this audio?", + "choice_a": "skip", + "choice_b": "please", + "choice_c": "start", + "choice_d": "continue reading", + "answer_gt": "skip", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7918 + }, + { + "path": "audio-1497883378.flac", + "question": "In this sound clip, which word related to 'house_place' is mentioned first?", + "choice_a": "and", + "choice_b": "the", + "choice_c": "bedroom", + "choice_d": "room", + "answer_gt": "bedroom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7919 + }, + { + "path": "audio-1498485762-headset.flac", + "question": "find the first word that's linked to 'joke_type' in this sound clip?", + "choice_a": "joke", + "choice_b": "me", + "choice_c": "baseball", + "choice_d": "funny", + "answer_gt": "baseball", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7920 + }, + { + "path": "audio-1499696298.flac", + "question": "Which word that signifies 'device_type' is spoken first in this clip?", + "choice_a": "off", + "choice_b": "i", + "choice_c": "vacuum", + "choice_d": "smart socket", + "answer_gt": "smart socket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7921 + }, + { + "path": "audio--1504194437.flac", + "question": "name the first word that's related to 'time' in this audio?", + "choice_a": "three", + "choice_b": "disable", + "choice_c": "three pm", + "choice_d": "five pm", + "answer_gt": "three pm", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7922 + }, + { + "path": "audio-1498570765-headset.flac", + "question": "What's the first word that's connected to 'date' in this audio?", + "choice_a": "the", + "choice_b": "is", + "choice_c": "the twenty third of this month", + "choice_d": "ten day", + "answer_gt": "ten day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7923 + }, + { + "path": "audio-1502216654-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'person'?", + "choice_a": "priyanka chopra", + "choice_b": "priyanka", + "choice_c": "amy", + "choice_d": "pawel's", + "answer_gt": "priyanka chopra", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7924 + }, + { + "path": "audio-1502101043.flac", + "question": "What's the first word that's tied to 'media_type' mentioned in this audio?", + "choice_a": "predictions", + "choice_b": "the", + "choice_c": "newest provider", + "choice_d": "bbc poll", + "answer_gt": "bbc poll", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7925 + }, + { + "path": "audio-1490007125-headset.flac", + "question": "tell me the first 'event_name'-connected word in this audio.", + "choice_a": "is", + "choice_b": "what", + "choice_c": "appointment", + "choice_d": "to have tablets", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7926 + }, + { + "path": "audio-1502376438.flac", + "question": "What's the earliest 'device_type'-related term you hear in this clip?", + "choice_a": "plug", + "choice_b": "wemo plug socket", + "choice_c": "robot vacuum cleaner", + "choice_d": "all the lights", + "answer_gt": "wemo plug socket", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7927 + }, + { + "path": "audio-1497879813-headset.flac", + "question": "What begins the list of 'date' words in this audio clip?", + "choice_a": "when", + "choice_b": "is", + "choice_c": "the twenty third of this month", + "choice_d": "friday", + "answer_gt": "friday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7928 + }, + { + "path": "audio-1502300210.flac", + "question": "What is the first word that's associated with 'person' in this audio?", + "choice_a": "you", + "choice_b": "to", + "choice_c": "john", + "choice_d": "sarah", + "answer_gt": "john", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7929 + }, + { + "path": "audio-1490108508-headset.flac", + "question": "What's the first word mentioned in this audio that's about 'event_name'?", + "choice_a": "everyone", + "choice_b": "for", + "choice_c": "birthday", + "choice_d": "concerts", + "answer_gt": "birthday", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7930 + }, + { + "path": "audio-1502200807-headset.flac", + "question": "What's the first word that's associated with 'person' in this clip?", + "choice_a": "jessica's", + "choice_b": "birthday", + "choice_c": "petey's", + "choice_d": "john smith", + "answer_gt": "jessica's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7931 + }, + { + "path": "audio-1501756067.flac", + "question": "What's the first word that's characteristic of 'definition_word' in this clip?", + "choice_a": "rock sand", + "choice_b": "describe", + "choice_c": "sphere", + "choice_d": "motivation", + "answer_gt": "rock sand", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7932 + }, + { + "path": "audio-1499360982.flac", + "question": "What's the first word that's alluding to 'relation' in this sound recording?", + "choice_a": "have", + "choice_b": "co worker", + "choice_c": "grandma", + "choice_d": "mom", + "answer_gt": "co worker", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7933 + }, + { + "path": "audio-1488985919.flac", + "question": "What's the first word that's indicative of 'business_type' in this audio?", + "choice_a": "can", + "choice_b": "movie", + "choice_c": "stores", + "choice_d": "furniture store", + "answer_gt": "movie", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7934 + }, + { + "path": "audio-1502311437.flac", + "question": "Which word comes first in this audio that's tied to 'person'?", + "choice_a": "rahul", + "choice_b": "marriage", + "choice_c": "amy", + "choice_d": "jeffrey burnette", + "answer_gt": "rahul", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7935 + }, + { + "path": "audio-1500036829-headset.flac", + "question": "In this audio clip, what's the first word that pertains to 'food_type'?", + "choice_a": "best", + "choice_b": "the", + "choice_c": "junk food", + "choice_d": "spaghetti", + "answer_gt": "spaghetti", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7936 + }, + { + "path": "audio-1490008561-headset.flac", + "question": "Which word comes first in this audio that's tied to 'event_name'?", + "choice_a": "a", + "choice_b": "function", + "choice_c": "meetings", + "choice_d": "call mom", + "answer_gt": "function", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7937 + }, + { + "path": "audio-1501758864-headset.flac", + "question": "What's the premier word about 'place_name' in this sound clip?", + "choice_a": "beijing", + "choice_b": "is", + "choice_c": "detroit", + "choice_d": "russia", + "answer_gt": "beijing", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7938 + }, + { + "path": "audio-1488969797.flac", + "question": "tell me the first word that's in relation to 'date' in this audio?", + "choice_a": "the", + "choice_b": "today", + "choice_c": "monday", + "choice_d": "fifth", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7939 + }, + { + "path": "audio--1505407266.flac", + "question": "What's the first word that's characteristic of 'player_setting' in this clip?", + "choice_a": "the", + "choice_b": "pause", + "choice_c": "replay", + "choice_d": "stop", + "answer_gt": "pause", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7940 + }, + { + "path": "audio-1495733835.flac", + "question": "What's the first word that's tied to 'person' mentioned in this audio?", + "choice_a": "two", + "choice_b": "jane's", + "choice_c": "jeffrey burnette", + "choice_d": "ben's", + "answer_gt": "jane's", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7941 + }, + { + "path": "audio-1490967708-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'device_type'?", + "choice_a": "the", + "choice_b": "all", + "choice_c": "all the lights", + "choice_d": "vacuum", + "answer_gt": "all the lights", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7942 + }, + { + "path": "audio-1490355033-headset.flac", + "question": "In this audio, what's the first word that's reminiscent of 'transport_type'?", + "choice_a": "train", + "choice_b": "to", + "choice_c": "metro", + "choice_d": "yellow cab", + "answer_gt": "train", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7943 + }, + { + "path": "audio-1499691666.flac", + "question": "tell me the first 'event_name'-connected word in this audio.", + "choice_a": "erase", + "choice_b": "appointment", + "choice_c": "dentist appointment", + "choice_d": "doctor appointment", + "answer_gt": "appointment", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7944 + }, + { + "path": "audio-1499692525.flac", + "question": "What's the first word that's characteristic of 'event_name' in this clip?", + "choice_a": "call", + "choice_b": "me", + "choice_c": "call mom", + "choice_d": "function", + "answer_gt": "call mom", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7945 + }, + { + "path": "audio--1504190472.flac", + "question": "Which word comes first in this audio that's tied to 'news_topic'?", + "choice_a": "the", + "choice_b": "election", + "choice_c": "international women's day", + "choice_d": "stock price for costco", + "answer_gt": "election", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7946 + }, + { + "path": "audio-1494512477-headset.flac", + "question": "What's the first word in this audio that has a connection to 'general_frequency'?", + "choice_a": "to", + "choice_b": "every", + "choice_c": "every day", + "choice_d": "hourly", + "answer_gt": "every day", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7947 + }, + { + "path": "audio-1501767336.flac", + "question": "What's the first word that's in association with 'player_setting' in this recording?", + "choice_a": "reading", + "choice_b": "please", + "choice_c": "playback", + "choice_d": "continue reading", + "answer_gt": "continue reading", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7948 + }, + { + "path": "audio-1497437209.flac", + "question": "What is the initial 'date'-associated word you hear in this recording?", + "choice_a": "day", + "choice_b": "august", + "choice_c": "one thousand nine hundred and ninety", + "choice_d": "august fifteenth", + "answer_gt": "august fifteenth", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7949 + }, + { + "path": "audio-1498572758-headset.flac", + "question": "What's the earliest 'weather_descriptor'-related term you hear in this clip?", + "choice_a": "temperature", + "choice_b": "what", + "choice_c": "raincoat", + "choice_d": "snow tyres", + "answer_gt": "temperature", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7950 + }, + { + "path": "audio-1501151388.flac", + "question": "detect the first mention of a word related to 'date' in this audio.", + "choice_a": "please", + "choice_b": "have", + "choice_c": "tomorrow", + "choice_d": "todays", + "answer_gt": "tomorrow", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7951 + }, + { + "path": "audio-1501752463.flac", + "question": "What's the first word that's tied to 'event_name' mentioned in this audio?", + "choice_a": "meeting", + "choice_b": "myself", + "choice_c": "fringe", + "choice_d": "pick up mark", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7952 + }, + { + "path": "audio-1500978977.flac", + "question": "Which word linked to 'place_name' comes first in this recording?", + "choice_a": "fiji", + "choice_b": "is", + "choice_c": "this city", + "choice_d": "one hundred and one", + "answer_gt": "fiji", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7953 + }, + { + "path": "audio-1501775574-headset.flac", + "question": "What's the first word that's representative of 'person' in this recording?", + "choice_a": "add", + "choice_b": "list", + "choice_c": "kate's", + "choice_d": "bob", + "answer_gt": "bob", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7954 + }, + { + "path": "audio-1501406713.flac", + "question": "What word relevant to 'transport_type' is mentioned at the start of this audio?", + "choice_a": "yellow", + "choice_b": "call", + "choice_c": "uberpool", + "choice_d": "yellow cab", + "answer_gt": "yellow cab", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7955 + }, + { + "path": "audio-1488984501.flac", + "question": "What word relevant to 'place_name' is mentioned at the start of this audio?", + "choice_a": "on", + "choice_b": "downtown", + "choice_c": "international", + "choice_d": "this city", + "answer_gt": "downtown", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7956 + }, + { + "path": "audio-1501772353.flac", + "question": "What's the first word that's tied to 'person' mentioned in this audio?", + "choice_a": "i", + "choice_b": "emails", + "choice_c": "barack obama", + "choice_d": "dan", + "answer_gt": "dan", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7957 + }, + { + "path": "audio-1502101993.flac", + "question": "What is the first word in this clip that's relevant to 'place_name'?", + "choice_a": "the", + "choice_b": "china", + "choice_c": "international", + "choice_d": "miami", + "answer_gt": "china", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7958 + }, + { + "path": "audio-1495734598.flac", + "question": "What's the first word mentioned in this audio that's about 'transport_type'?", + "choice_a": "at", + "choice_b": "taxi", + "choice_c": "taxi", + "choice_d": "train", + "answer_gt": "taxi", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7959 + }, + { + "path": "audio-1502895701-headset.flac", + "question": "What's the first word that's alluding to 'color_type' in this sound recording?", + "choice_a": "is", + "choice_b": "pink", + "choice_c": "red ish", + "choice_d": "blue", + "answer_gt": "pink", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7960 + }, + { + "path": "audio-1501693693.flac", + "question": "In this sound clip, which word related to 'media_type' is mentioned first?", + "choice_a": "tweet", + "choice_b": "about", + "choice_c": "bob's news", + "choice_d": "so clean", + "answer_gt": "tweet", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7961 + }, + { + "path": "audio-1499267168.flac", + "question": "detect the first mention of a word related to 'date' in this audio.", + "choice_a": "to", + "choice_b": "the", + "choice_c": "this week", + "choice_d": "this Friday", + "answer_gt": "this week", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7962 + }, + { + "path": "audio-1502101093-headset.flac", + "question": "In this sound clip, what's the first word that's relevant to 'business_name'?", + "choice_a": "@companyname", + "choice_b": "at", + "choice_c": "company name", + "choice_d": "terranova", + "answer_gt": "@companyname", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7963 + }, + { + "path": "audio-1502101733.flac", + "question": "What's the first word that's indicative of 'currency_name' in this audio?", + "choice_a": "dollar", + "choice_b": "at", + "choice_c": "dolla", + "choice_d": "dollar", + "answer_gt": "dollar", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7964 + }, + { + "path": "audio-1498563804-headset.flac", + "question": "What is the initial 'radio_name'-associated word you hear in this recording?", + "choice_a": "xm", + "choice_b": "start", + "choice_c": "bbc", + "choice_d": "sirius xm radio channel", + "answer_gt": "sirius xm radio channel", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7965 + }, + { + "path": "audio-1495734915-headset.flac", + "question": "identify the initial word associated with 'person' in this audio.", + "choice_a": "did", + "choice_b": "george clooney", + "choice_c": "trump", + "choice_d": "naty", + "answer_gt": "george clooney", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7966 + }, + { + "path": "audio-1501772302.flac", + "question": "tell me the first 'podcast_name'-connected word in this audio.", + "choice_a": "eagles", + "choice_b": "philadelphia", + "choice_c": "philadelphia eagles content", + "choice_d": "disney", + "answer_gt": "philadelphia eagles content", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7967 + }, + { + "path": "audio-1498484148.flac", + "question": "What's the first word that's representative of 'time' in this recording?", + "choice_a": "last", + "choice_b": "ten", + "choice_c": "sunrise", + "choice_d": "last ten minutes", + "answer_gt": "last ten minutes", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7968 + }, + { + "path": "audio-1501767950-headset.flac", + "question": "What is the initial 'event_name'-associated word you hear in this recording?", + "choice_a": "meeting", + "choice_b": "put", + "choice_c": "meeting", + "choice_d": "pick up mark", + "answer_gt": "meeting", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7969 + }, + { + "path": "audio-1502896214.flac", + "question": "Which word that's affiliated with 'date' is spoken first in this clip?", + "choice_a": "today", + "choice_b": "schedule", + "choice_c": "this week's", + "choice_d": "yesterday", + "answer_gt": "today", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7970 + }, + { + "path": "audio-1502376386-headset.flac", + "question": "What's the first word that's in association with 'event_name' in this recording?", + "choice_a": "next", + "choice_b": "fringe", + "choice_c": "second of next mont lands", + "choice_d": "meetings", + "answer_gt": "fringe", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7971 + }, + { + "path": "audio-1502376916.flac", + "question": "In this sound clip, what's the first word that's relevant to 'game_name'?", + "choice_a": "and", + "choice_b": "begin", + "choice_c": "harry potter and the chamber of secrets", + "choice_d": "chess", + "answer_gt": "chess", + "task_name": "Speech_Entity_Reconition", + "dataset_name": "slurp", + "uniq_id": 7972 + }, + { + "path": "audio-1495734419.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "hue_lightoff", + "choice_b": "social_query", + "choice_c": "general_quirky", + "choice_d": "transport_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7973 + }, + { + "path": "audio-1499691888.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "play_music", + "choice_b": "joke", + "choice_c": "iot_hue_lightchange", + "choice_d": "email_querycontact", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7974 + }, + { + "path": "audio-1502377451.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "news_query", + "choice_b": "play_audiobook", + "choice_c": "post", + "choice_d": "sendemail", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7975 + }, + { + "path": "audio-1490289532.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "calendar_remove", + "choice_b": "datetime_query", + "choice_c": "qa_definition", + "choice_d": "audio_volume_other", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7976 + }, + { + "path": "audio-1501776383-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "general_quirky", + "choice_b": "recommendation_events", + "choice_c": "transport_traffic", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7977 + }, + { + "path": "audio-1498571093-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "calendar_query", + "choice_b": "remove", + "choice_c": "iot_hue_lighton", + "choice_d": "iot_coffee", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7978 + }, + { + "path": "audio-1490195239.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "recommendation_locations", + "choice_c": "calendar_set", + "choice_d": "music", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7979 + }, + { + "path": "audio-1434535625.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "music_likeness", + "choice_b": "query", + "choice_c": "datetime_query", + "choice_d": "play_game", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7980 + }, + { + "path": "audio-1489154983.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "qa_stock", + "choice_b": "qa_currency", + "choice_c": "iot_cleaning", + "choice_d": "play_podcasts", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7981 + }, + { + "path": "audio-1489153247.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "audio_volume_up", + "choice_b": "news_query", + "choice_c": "lists_createoradd", + "choice_d": "play_podcasts", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7982 + }, + { + "path": "audio-1498566599.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "iot_wemo_on", + "choice_b": "transport_ticket", + "choice_c": "calendar_remove", + "choice_d": "qa_currency", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7983 + }, + { + "path": "audio-1490109978-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "weather_query", + "choice_b": "music_settings", + "choice_c": "qa_factoid", + "choice_d": "music", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7984 + }, + { + "path": "audio-1497865976.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "joke", + "choice_b": "qa_factoid", + "choice_c": "iot_hue_lightdim", + "choice_d": "transport_taxi", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7985 + }, + { + "path": "audio-1498564763.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "joke", + "choice_b": "iot_cleaning", + "choice_c": "iot_hue_lightdim", + "choice_d": "email_addcontact", + "answer_gt": "joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7986 + }, + { + "path": "audio-1501753559-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "social_post", + "choice_b": "qa_stock", + "choice_c": "transport_ticket", + "choice_d": "lists_createoradd", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7987 + }, + { + "path": "audio-1494418635.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "calendar_set", + "choice_c": "query", + "choice_d": "datetime_convert", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7988 + }, + { + "path": "audio-1501687757-headset.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "qa_maths", + "choice_b": "general_quirky", + "choice_c": "datetime_query", + "choice_d": "general_greet", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7989 + }, + { + "path": "audio-1499265306-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "email_query", + "choice_b": "play_game", + "choice_c": "recommendation_events", + "choice_d": "iot_wemo_on", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7990 + }, + { + "path": "audio-1500899982.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "takeaway_query", + "choice_b": "play_game", + "choice_c": "social_post", + "choice_d": "qa_currency", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7991 + }, + { + "path": "audio-1489154114.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "lists_remove", + "choice_b": "post", + "choice_c": "sendemail", + "choice_d": "music_likeness", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7992 + }, + { + "path": "audio-1502111927-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "audio_volume_mute", + "choice_c": "iot_hue_lightchange", + "choice_d": "play_podcasts", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7993 + }, + { + "path": "audio-1501778696.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "iot_hue_lightup", + "choice_b": "audio_volume_down", + "choice_c": "iot_hue_lightchange", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7994 + }, + { + "path": "audio-1490367273.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "datetime_convert", + "choice_b": "iot_hue_lightup", + "choice_c": "music_query", + "choice_d": "lists_createoradd", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7995 + }, + { + "path": "audio--1504197871-headset.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "hue_lightdim", + "choice_b": "alarm_set", + "choice_c": "calendar_set", + "choice_d": "iot_wemo_off", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7996 + }, + { + "path": "audio-1502375028.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "email_addcontact", + "choice_b": "calendar_remove", + "choice_c": "play_podcasts", + "choice_d": "music_likeness", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7997 + }, + { + "path": "audio-1502375973-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "lists_createoradd", + "choice_c": "transport_ticket", + "choice_d": "calendar_set", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7998 + }, + { + "path": "audio-1495376449-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "iot_wemo_off", + "choice_b": "locations", + "choice_c": "recommendation_locations", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 7999 + }, + { + "path": "audio-1490107043-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "lists_createoradd", + "choice_b": "email_sendemail", + "choice_c": "transport_taxi", + "choice_d": "iot_coffee", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8000 + }, + { + "path": "audio-1498646746.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "qa_definition", + "choice_b": "calendar_query", + "choice_c": "audio_volume_up", + "choice_d": "transport_traffic", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8001 + }, + { + "path": "audio-1434528443-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "datetime_query", + "choice_b": "play_music", + "choice_c": "calendar_remove", + "choice_d": "iot_cleaning", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8002 + }, + { + "path": "audio-1501690106.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "iot_coffee", + "choice_b": "general_quirky", + "choice_c": "lists_query", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8003 + }, + { + "path": "audio-1499360330.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "iot_hue_lightchange", + "choice_b": "email_querycontact", + "choice_c": "alarm_remove", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8004 + }, + { + "path": "audio--1504198760.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "lists_query", + "choice_b": "iot_wemo_off", + "choice_c": "alarm_set", + "choice_d": "music_likeness", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8005 + }, + { + "path": "audio-1434535581.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "music_query", + "choice_b": "lists_query", + "choice_c": "iot_hue_lightdim", + "choice_d": "recommendation_movies", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8006 + }, + { + "path": "audio-1499266775.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "email_addcontact", + "choice_b": "weather_query", + "choice_c": "audio_volume_up", + "choice_d": "post", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8007 + }, + { + "path": "audio-1434541421-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "transport_query", + "choice_b": "email_sendemail", + "choice_c": "social_post", + "choice_d": "music_query", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8008 + }, + { + "path": "audio-1490704103-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "transport_traffic", + "choice_b": "audio_volume_down", + "choice_c": "calendar_set", + "choice_d": "iot_hue_lightup", + "answer_gt": "audio_volume_down", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8009 + }, + { + "path": "audio-1488991837.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "cooking_recipe", + "choice_b": "takeaway_order", + "choice_c": "transport_ticket", + "choice_d": "weather_query", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8010 + }, + { + "path": "audio-1502377176-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "hue_lightdim", + "choice_b": "general_greet", + "choice_c": "play_radio", + "choice_d": "calendar_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8011 + }, + { + "path": "audio-1490705108-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "iot_hue_lightoff", + "choice_b": "iot_hue_lighton", + "choice_c": "query", + "choice_d": "iot_hue_lightup", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8012 + }, + { + "path": "audio-1501778149.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "takeaway_query", + "choice_b": "social_post", + "choice_c": "transport_ticket", + "choice_d": "joke", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8013 + }, + { + "path": "audio-1500039567-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_query", + "choice_b": "transport_taxi", + "choice_c": "iot_cleaning", + "choice_d": "iot_coffee", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8014 + }, + { + "path": "audio-1497556544-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "email_querycontact", + "choice_b": "email_sendemail", + "choice_c": "music_dislikeness", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8015 + }, + { + "path": "audio-1497883730.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "iot_hue_lighton", + "choice_b": "music_settings", + "choice_c": "datetime_query", + "choice_d": "transport_traffic", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8016 + }, + { + "path": "audio-1501152552.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "music_query", + "choice_b": "datetime_query", + "choice_c": "music_settings", + "choice_d": "play_audiobook", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8017 + }, + { + "path": "audio-1497495664-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "calendar_query", + "choice_b": "weather_query", + "choice_c": "iot_hue_lightdim", + "choice_d": "qa_maths", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8018 + }, + { + "path": "audio-1488975923.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "audio_volume_down", + "choice_b": "general_quirky", + "choice_c": "recommendation_events", + "choice_d": "takeaway_order", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8019 + }, + { + "path": "audio--1504198651-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "iot_coffee", + "choice_b": "audio_volume_other", + "choice_c": "music_dislikeness", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8020 + }, + { + "path": "audio--1504190994-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "news_query", + "choice_b": "play_music", + "choice_c": "takeaway_order", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8021 + }, + { + "path": "audio-1488558888.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "lists_query", + "choice_b": "takeaway_order", + "choice_c": "music_query", + "choice_d": "play_audiobook", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8022 + }, + { + "path": "audio--1504198594.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "hue_lightdim", + "choice_b": "music_query", + "choice_c": "alarm_query", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8023 + }, + { + "path": "audio-1497555556.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "news_query", + "choice_b": "calendar_query", + "choice_c": "music_likeness", + "choice_d": "iot_hue_lightoff", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8024 + }, + { + "path": "audio-1497192547.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "audio_volume_up", + "choice_b": "qa_definition", + "choice_c": "lists_remove", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8025 + }, + { + "path": "audio-1489497168-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "audio_volume_other", + "choice_b": "iot_hue_lightoff", + "choice_c": "transport_traffic", + "choice_d": "recommendation_events", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8026 + }, + { + "path": "audio-1501758071-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "general_joke", + "choice_b": "lists_createoradd", + "choice_c": "recommendation_events", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8027 + }, + { + "path": "audio-1501688180.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "email_sendemail", + "choice_b": "iot_wemo_on", + "choice_c": "locations", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8028 + }, + { + "path": "audio-1499697575.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "alarm_query", + "choice_b": "qa_definition", + "choice_c": "social_query", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8029 + }, + { + "path": "audio-1501691356-headset.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "weather_query", + "choice_b": "transport_taxi", + "choice_c": "sendemail", + "choice_d": "datetime_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8030 + }, + { + "path": "audio-1490103360.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "iot_hue_lightoff", + "choice_b": "createoradd", + "choice_c": "audio_volume_other", + "choice_d": "weather_query", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8031 + }, + { + "path": "audio-1501758978.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "play_music", + "choice_b": "audio_volume_down", + "choice_c": "play_radio", + "choice_d": "iot_cleaning", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8032 + }, + { + "path": "audio-1490202467.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "music_likeness", + "choice_b": "locations", + "choice_c": "news_query", + "choice_d": "joke", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8033 + }, + { + "path": "audio-1498576747.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "music_settings", + "choice_b": "qa_factoid", + "choice_c": "calendar_query", + "choice_d": "general_quirky", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8034 + }, + { + "path": "audio-1502198861.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "post", + "choice_b": "lists_createoradd", + "choice_c": "cooking_recipe", + "choice_d": "locations", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8035 + }, + { + "path": "audio-1490281122-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "hue_lightdim", + "choice_b": "calendar_query", + "choice_c": "play_audiobook", + "choice_d": "recommendation_movies", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8036 + }, + { + "path": "audio-1502376770-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_definition", + "choice_b": "play_podcasts", + "choice_c": "play_music", + "choice_d": "calendar_remove", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8037 + }, + { + "path": "audio-1502893679-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "datetime_query", + "choice_b": "qa_maths", + "choice_c": "email_addcontact", + "choice_d": "general_joke", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8038 + }, + { + "path": "audio-1434539939.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "iot_coffee", + "choice_b": "email_sendemail", + "choice_c": "recommendation_events", + "choice_d": "social_post", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8039 + }, + { + "path": "audio-1499089337-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "play_music", + "choice_b": "post", + "choice_c": "transport_ticket", + "choice_d": "email_querycontact", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8040 + }, + { + "path": "audio--1504198727-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "qa_maths", + "choice_b": "audio_volume_mute", + "choice_c": "calendar_query", + "choice_d": "takeaway_order", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8041 + }, + { + "path": "audio--1504192520.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "sendemail", + "choice_b": "play_music", + "choice_c": "locations", + "choice_d": "audio_volume_down", + "answer_gt": "audio_volume_down", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8042 + }, + { + "path": "audio--1506521206.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "email_addcontact", + "choice_b": "audio_volume_down", + "choice_c": "general_quirky", + "choice_d": "transport_traffic", + "answer_gt": "audio_volume_down", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8043 + }, + { + "path": "audio-1502302042-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "play_radio", + "choice_b": "audio_volume_up", + "choice_c": "alarm_query", + "choice_d": "iot_hue_lightchange", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8044 + }, + { + "path": "audio-1499697377.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "social_post", + "choice_b": "lists_remove", + "choice_c": "calendar_set", + "choice_d": "query", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8045 + }, + { + "path": "audio-1492785446.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "qa_factoid", + "choice_b": "sendemail", + "choice_c": "radio", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8046 + }, + { + "path": "audio-1490200090.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "weather_query", + "choice_b": "lists_remove", + "choice_c": "iot_hue_lighton", + "choice_d": "alarm_set", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8047 + }, + { + "path": "audio-1502893785.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "qa_currency", + "choice_b": "audio_volume_other", + "choice_c": "cooking_recipe", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8048 + }, + { + "path": "audio-1501688159.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "play_music", + "choice_b": "transport_traffic", + "choice_c": "general_quirky", + "choice_d": "iot_hue_lightoff", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8049 + }, + { + "path": "audio--1505324888.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "play_radio", + "choice_b": "recommendation_events", + "choice_c": "alarm_set", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8050 + }, + { + "path": "audio-1489672155-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "weather_query", + "choice_b": "lists_query", + "choice_c": "cooking_recipe", + "choice_d": "play_audiobook", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8051 + }, + { + "path": "audio-1498484871-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "alarm_remove", + "choice_b": "calendar_set", + "choice_c": "play_radio", + "choice_d": "news_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8052 + }, + { + "path": "audio-1498759071.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "iot_cleaning", + "choice_b": "music_likeness", + "choice_c": "play_audiobook", + "choice_d": "email_querycontact", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8053 + }, + { + "path": "audio--1506079542.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_traffic", + "choice_b": "qa_maths", + "choice_c": "play_audiobook", + "choice_d": "play_music", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8054 + }, + { + "path": "audio-1434544328.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "general_joke", + "choice_b": "general_quirky", + "choice_c": "cooking_recipe", + "choice_d": "sendemail", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8055 + }, + { + "path": "audio--1504194760-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "hue_lightdim", + "choice_b": "audio_volume_mute", + "choice_c": "iot_hue_lightup", + "choice_d": "email_sendemail", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8056 + }, + { + "path": "audio-1501767992-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "play_radio", + "choice_b": "general_joke", + "choice_c": "hue_lightdim", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8057 + }, + { + "path": "audio-1501151545.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "general_quirky", + "choice_b": "calendar_query", + "choice_c": "qa_currency", + "choice_d": "music_likeness", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8058 + }, + { + "path": "audio-1497861661-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "transport_query", + "choice_b": "transport_taxi", + "choice_c": "play_game", + "choice_d": "recommendation_locations", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8059 + }, + { + "path": "audio--1505323813-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "email_addcontact", + "choice_b": "alarm_query", + "choice_c": "locations", + "choice_d": "qa_definition", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8060 + }, + { + "path": "audio-1498568891-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "recommendation_movies", + "choice_b": "calendar_set", + "choice_c": "qa_stock", + "choice_d": "lists_createoradd", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8061 + }, + { + "path": "audio-1434538853.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "music_settings", + "choice_b": "takeaway_query", + "choice_c": "email_querycontact", + "choice_d": "play_music", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8062 + }, + { + "path": "audio-1488984587.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "music_query", + "choice_b": "iot_wemo_off", + "choice_c": "transport_traffic", + "choice_d": "iot_coffee", + "answer_gt": "transport_traffic", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8063 + }, + { + "path": "audio-1502216864-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "datetime_convert", + "choice_b": "email_sendemail", + "choice_c": "email_query", + "choice_d": "alarm_set", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8064 + }, + { + "path": "audio-1502890146-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "social_query", + "choice_b": "music_query", + "choice_c": "email_sendemail", + "choice_d": "datetime_query", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8065 + }, + { + "path": "audio-1499360716.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "query", + "choice_b": "iot_wemo_on", + "choice_c": "play_audiobook", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8066 + }, + { + "path": "audio-1499691643-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "iot_coffee", + "choice_b": "qa_currency", + "choice_c": "audio_volume_down", + "choice_d": "hue_lightoff", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8067 + }, + { + "path": "audio-1489497544-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "general_greet", + "choice_b": "iot_hue_lighton", + "choice_c": "email_querycontact", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8068 + }, + { + "path": "audio-1501690119-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "iot_wemo_off", + "choice_b": "iot_hue_lightchange", + "choice_c": "iot_hue_lighton", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8069 + }, + { + "path": "audio-1501772591-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "recommendation_movies", + "choice_b": "audio_volume_up", + "choice_c": "news_query", + "choice_d": "iot_hue_lightup", + "answer_gt": "audio_volume_up", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8070 + }, + { + "path": "audio-1488985787.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "weather_query", + "choice_b": "radio", + "choice_c": "query", + "choice_d": "lists_remove", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8071 + }, + { + "path": "audio-1492784627.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "general_joke", + "choice_b": "lists_remove", + "choice_c": "hue_lightdim", + "choice_d": "iot_hue_lightdim", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8072 + }, + { + "path": "audio-1502309987-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "qa_factoid", + "choice_b": "iot_wemo_off", + "choice_c": "qa_maths", + "choice_d": "post", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8073 + }, + { + "path": "audio-1492783811.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "play_audiobook", + "choice_b": "iot_hue_lightup", + "choice_c": "music_settings", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8074 + }, + { + "path": "audio-1499692212-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "transport_traffic", + "choice_b": "lists_query", + "choice_c": "audio_volume_up", + "choice_d": "social_post", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8075 + }, + { + "path": "audio-1501697209.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "recommendation_events", + "choice_b": "lists_query", + "choice_c": "qa_currency", + "choice_d": "takeaway_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8076 + }, + { + "path": "audio-1497861648.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "music_query", + "choice_b": "email_query", + "choice_c": "general_quirky", + "choice_d": "iot_hue_lightchange", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8077 + }, + { + "path": "audio-1502376242-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "datetime_query", + "choice_b": "recommendation_movies", + "choice_c": "qa_stock", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8078 + }, + { + "path": "audio-1494419952.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "calendar_query", + "choice_b": "email_addcontact", + "choice_c": "recommendation_locations", + "choice_d": "qa_currency", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8079 + }, + { + "path": "audio-1490292188.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "general_joke", + "choice_b": "calendar_set", + "choice_c": "qa_definition", + "choice_d": "cooking_recipe", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8080 + }, + { + "path": "audio-1501688284.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "news_query", + "choice_b": "qa_currency", + "choice_c": "cooking_recipe", + "choice_d": "hue_lightdim", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8081 + }, + { + "path": "audio-1502375627-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "iot_hue_lightchange", + "choice_b": "email_addcontact", + "choice_c": "hue_lightoff", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8082 + }, + { + "path": "audio--1504191051-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "lists_createoradd", + "choice_b": "play_audiobook", + "choice_c": "iot_hue_lightchange", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8083 + }, + { + "path": "audio-1490706286-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "general_joke", + "choice_b": "play_audiobook", + "choice_c": "email_addcontact", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8084 + }, + { + "path": "audio-1501694842.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "social_query", + "choice_b": "weather_query", + "choice_c": "music_query", + "choice_d": "iot_hue_lightup", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8085 + }, + { + "path": "audio-1501764360-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "sendemail", + "choice_b": "iot_hue_lightchange", + "choice_c": "play_audiobook", + "choice_d": "iot_hue_lightup", + "answer_gt": "iot_hue_lightup", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8086 + }, + { + "path": "audio-1490198338-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "qa_currency", + "choice_b": "lists_remove", + "choice_c": "alarm_set", + "choice_d": "email_addcontact", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8087 + }, + { + "path": "audio-1501690561.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "general_quirky", + "choice_b": "social_query", + "choice_c": "music_settings", + "choice_d": "remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8088 + }, + { + "path": "audio-1502100921-headset.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "music_settings", + "choice_b": "takeaway_query", + "choice_c": "general_joke", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8089 + }, + { + "path": "audio-1498572936.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "recommendation_movies", + "choice_b": "play_music", + "choice_c": "post", + "choice_d": "radio", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8090 + }, + { + "path": "audio-1490183698.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "audio_volume_other", + "choice_b": "play_music", + "choice_c": "play_game", + "choice_d": "lists_remove", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8091 + }, + { + "path": "audio-1490957508.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "audio_volume_other", + "choice_b": "music_query", + "choice_c": "datetime_convert", + "choice_d": "qa_factoid", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8092 + }, + { + "path": "audio-1496265876-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "general_greet", + "choice_b": "calendar_set", + "choice_c": "cooking_recipe", + "choice_d": "email_querycontact", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8093 + }, + { + "path": "audio-1502199914.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "music_settings", + "choice_b": "music_likeness", + "choice_c": "news_query", + "choice_d": "recommendation_movies", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8094 + }, + { + "path": "audio-1501685732-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "news_query", + "choice_b": "takeaway_order", + "choice_c": "iot_hue_lighton", + "choice_d": "audio_volume_other", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8095 + }, + { + "path": "audio-1497864565-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "radio", + "choice_b": "social_query", + "choice_c": "takeaway_order", + "choice_d": "lists_createoradd", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8096 + }, + { + "path": "audio-1502199985-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "post", + "choice_b": "audio_volume_mute", + "choice_c": "qa_factoid", + "choice_d": "createoradd", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8097 + }, + { + "path": "audio-1495369607-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "play_music", + "choice_b": "iot_hue_lightdim", + "choice_c": "lists_query", + "choice_d": "email_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8098 + }, + { + "path": "audio-1502200460-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "iot_hue_lightdim", + "choice_b": "alarm_set", + "choice_c": "datetime_convert", + "choice_d": "takeaway_query", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8099 + }, + { + "path": "audio-1497024791.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "transport_taxi", + "choice_b": "takeaway_query", + "choice_c": "lists_createoradd", + "choice_d": "createoradd", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8100 + }, + { + "path": "audio-1488992846.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "transport_taxi", + "choice_b": "query", + "choice_c": "hue_lightdim", + "choice_d": "music_query", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8101 + }, + { + "path": "audio-1490181750-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "general_quirky", + "choice_b": "email_sendemail", + "choice_c": "alarm_remove", + "choice_d": "post", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8102 + }, + { + "path": "audio-1490021944-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "general_quirky", + "choice_b": "transport_query", + "choice_c": "audio_volume_up", + "choice_d": "email_addcontact", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8103 + }, + { + "path": "audio-1502218791.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "iot_wemo_on", + "choice_b": "social_post", + "choice_c": "audio_volume_mute", + "choice_d": "lists_query", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8104 + }, + { + "path": "audio-1501687705.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "lists_query", + "choice_b": "hue_lightdim", + "choice_c": "alarm_query", + "choice_d": "transport_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8105 + }, + { + "path": "audio-1488638656.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "social_query", + "choice_b": "hue_lightoff", + "choice_c": "news_query", + "choice_d": "play_music", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8106 + }, + { + "path": "audio--1504194259.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "email_addcontact", + "choice_b": "alarm_set", + "choice_c": "audio_volume_up", + "choice_d": "general_joke", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8107 + }, + { + "path": "audio-1434533222-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "social_post", + "choice_b": "calendar_set", + "choice_c": "recommendation_locations", + "choice_d": "lists_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8108 + }, + { + "path": "audio-1489152700.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "iot_cleaning", + "choice_b": "email_sendemail", + "choice_c": "transport_ticket", + "choice_d": "audio_volume_other", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8109 + }, + { + "path": "audio-1490203063.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "hue_lightoff", + "choice_b": "locations", + "choice_c": "calendar_remove", + "choice_d": "alarm_set", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8110 + }, + { + "path": "audio-1502101423-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "general_quirky", + "choice_b": "lists_createoradd", + "choice_c": "iot_hue_lightchange", + "choice_d": "general_joke", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8111 + }, + { + "path": "audio-1502896042.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "qa_definition", + "choice_c": "calendar_set", + "choice_d": "recommendation_locations", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8112 + }, + { + "path": "audio-1434533567-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "email_querycontact", + "choice_b": "news_query", + "choice_c": "play_podcasts", + "choice_d": "qa_currency", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8113 + }, + { + "path": "audio-1497437771-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "alarm_set", + "choice_b": "weather_query", + "choice_c": "general_greet", + "choice_d": "calendar_set", + "answer_gt": "general_greet", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8114 + }, + { + "path": "audio-1501151963-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "music_settings", + "choice_b": "query", + "choice_c": "iot_hue_lighton", + "choice_d": "email_querycontact", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8115 + }, + { + "path": "audio-1434532258-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "news_query", + "choice_b": "email_query", + "choice_c": "joke", + "choice_d": "social_post", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8116 + }, + { + "path": "audio-1490705843.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "radio", + "choice_b": "email_querycontact", + "choice_c": "news_query", + "choice_d": "qa_maths", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8117 + }, + { + "path": "audio-1501415524-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "email_querycontact", + "choice_b": "email_addcontact", + "choice_c": "play_audiobook", + "choice_d": "iot_wemo_on", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8118 + }, + { + "path": "audio-1490007028-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "qa_maths", + "choice_b": "takeaway_query", + "choice_c": "takeaway_order", + "choice_d": "play_game", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8119 + }, + { + "path": "audio-1501691380.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "lists_query", + "choice_b": "transport_ticket", + "choice_c": "email_querycontact", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8120 + }, + { + "path": "audio-1490704220-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "social_query", + "choice_b": "hue_lightdim", + "choice_c": "email_querycontact", + "choice_d": "takeaway_query", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8121 + }, + { + "path": "audio-1498485487-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "lists_createoradd", + "choice_b": "weather_query", + "choice_c": "iot_coffee", + "choice_d": "iot_wemo_on", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8122 + }, + { + "path": "audio-1501775683-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "remove", + "choice_b": "general_quirky", + "choice_c": "transport_taxi", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8123 + }, + { + "path": "audio-1494512423.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "weather_query", + "choice_b": "audio_volume_mute", + "choice_c": "play_podcasts", + "choice_d": "iot_hue_lightdim", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8124 + }, + { + "path": "audio--1505405963-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "calendar_query", + "choice_b": "play_audiobook", + "choice_c": "play_radio", + "choice_d": "datetime_convert", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8125 + }, + { + "path": "audio-1494512705-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "iot_hue_lightoff", + "choice_b": "cooking_recipe", + "choice_c": "weather_query", + "choice_d": "recommendation_locations", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8126 + }, + { + "path": "audio-1494607536-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "audio_volume_mute", + "choice_b": "takeaway_query", + "choice_c": "music_query", + "choice_d": "social_post", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8127 + }, + { + "path": "audio-1502099647-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "email_sendemail", + "choice_c": "qa_currency", + "choice_d": "transport_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8128 + }, + { + "path": "audio-1490194460.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "qa_currency", + "choice_b": "play_radio", + "choice_c": "iot_wemo_off", + "choice_d": "alarm_query", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8129 + }, + { + "path": "audio-1499692773.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "createoradd", + "choice_b": "datetime_query", + "choice_c": "sendemail", + "choice_d": "email_addcontact", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8130 + }, + { + "path": "audio-1501767740.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "email_query", + "choice_b": "email_addcontact", + "choice_c": "recommendation_events", + "choice_d": "transport_taxi", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8131 + }, + { + "path": "audio-1490703761.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "general_quirky", + "choice_b": "hue_lightoff", + "choice_c": "iot_hue_lightdim", + "choice_d": "calendar_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8132 + }, + { + "path": "audio-1492784264-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "weather_query", + "choice_c": "audio_volume_other", + "choice_d": "lists_createoradd", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8133 + }, + { + "path": "audio-1502376672-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "takeaway_query", + "choice_b": "email_sendemail", + "choice_c": "calendar_query", + "choice_d": "iot_hue_lightdim", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8134 + }, + { + "path": "audio-1490007546-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "play_podcasts", + "choice_b": "general_joke", + "choice_c": "lists_remove", + "choice_d": "transport_ticket", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8135 + }, + { + "path": "audio-1489498578-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "general_greet", + "choice_b": "general_quirky", + "choice_c": "hue_lightdim", + "choice_d": "music_settings", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8136 + }, + { + "path": "audio-1501757126-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "weather_query", + "choice_c": "general_joke", + "choice_d": "news_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8137 + }, + { + "path": "audio-1502200095-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "social_post", + "choice_b": "music_query", + "choice_c": "transport_traffic", + "choice_d": "audio_volume_up", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8138 + }, + { + "path": "audio-1501767336-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "lists_query", + "choice_b": "play_audiobook", + "choice_c": "datetime_query", + "choice_d": "social_query", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8139 + }, + { + "path": "audio--1505406815.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "email_sendemail", + "choice_b": "audio_volume_down", + "choice_c": "query", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8140 + }, + { + "path": "audio-1490367215.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "calendar_set", + "choice_b": "play_audiobook", + "choice_c": "takeaway_query", + "choice_d": "datetime_convert", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8141 + }, + { + "path": "audio-1488622278.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "general_joke", + "choice_b": "transport_ticket", + "choice_c": "weather_query", + "choice_d": "takeaway_order", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8142 + }, + { + "path": "audio-1500036086.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "play_radio", + "choice_b": "transport_query", + "choice_c": "qa_currency", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8143 + }, + { + "path": "audio-1502896210-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "music_dislikeness", + "choice_b": "calendar_query", + "choice_c": "iot_wemo_off", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8144 + }, + { + "path": "audio-1490366818-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "hue_lightdim", + "choice_b": "lists_createoradd", + "choice_c": "lists_query", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8145 + }, + { + "path": "audio-1502198907.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "recommendation_movies", + "choice_b": "calendar_set", + "choice_c": "recommendation_events", + "choice_d": "remove", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8146 + }, + { + "path": "audio-1501157834-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "takeaway_query", + "choice_b": "radio", + "choice_c": "play_game", + "choice_d": "music_settings", + "answer_gt": "music_settings", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8147 + }, + { + "path": "audio-1497862840-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "music_dislikeness", + "choice_b": "sendemail", + "choice_c": "datetime_query", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8148 + }, + { + "path": "audio-1501758993-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "iot_hue_lightoff", + "choice_b": "alarm_set", + "choice_c": "music_dislikeness", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8149 + }, + { + "path": "audio-1501687247.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "email_querycontact", + "choice_b": "alarm_query", + "choice_c": "calendar_query", + "choice_d": "alarm_remove", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8150 + }, + { + "path": "audio-1490357607.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "takeaway_query", + "choice_b": "hue_lightoff", + "choice_c": "audio_volume_down", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8151 + }, + { + "path": "audio-1501754231.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "iot_hue_lightchange", + "choice_c": "qa_maths", + "choice_d": "post", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8152 + }, + { + "path": "audio-1497616359-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "email_sendemail", + "choice_b": "locations", + "choice_c": "iot_coffee", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8153 + }, + { + "path": "audio-1497030563.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "audio_volume_other", + "choice_b": "email_addcontact", + "choice_c": "calendar_query", + "choice_d": "social_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8154 + }, + { + "path": "audio-1494606899-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "takeaway_order", + "choice_b": "iot_wemo_off", + "choice_c": "music_settings", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8155 + }, + { + "path": "audio-1492784451-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "iot_hue_lightoff", + "choice_b": "locations", + "choice_c": "createoradd", + "choice_d": "datetime_query", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8156 + }, + { + "path": "audio-1501686088-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "joke", + "choice_b": "lists_createoradd", + "choice_c": "weather_query", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8157 + }, + { + "path": "audio-1502309218.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "takeaway_order", + "choice_b": "social_post", + "choice_c": "lists_remove", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8158 + }, + { + "path": "audio-1497883739.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "iot_hue_lightup", + "choice_b": "general_quirky", + "choice_c": "music_settings", + "choice_d": "recommendation_movies", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8159 + }, + { + "path": "audio-1490203650.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "news_query", + "choice_b": "weather_query", + "choice_c": "alarm_query", + "choice_d": "cooking_recipe", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8160 + }, + { + "path": "audio-1498570380-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "email_querycontact", + "choice_b": "calendar_set", + "choice_c": "play_radio", + "choice_d": "iot_wemo_off", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8161 + }, + { + "path": "audio-1502374771-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "audio_volume_mute", + "choice_b": "alarm_remove", + "choice_c": "audio_volume_up", + "choice_d": "calendar_set", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8162 + }, + { + "path": "audio-1501758119-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "calendar_remove", + "choice_b": "play_podcasts", + "choice_c": "email_sendemail", + "choice_d": "recommendation_movies", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8163 + }, + { + "path": "audio-1497861959-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "play_podcasts", + "choice_b": "calendar_set", + "choice_c": "takeaway_order", + "choice_d": "radio", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8164 + }, + { + "path": "audio-1498572936-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "iot_wemo_on", + "choice_b": "play_music", + "choice_c": "music_query", + "choice_d": "weather_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8165 + }, + { + "path": "audio-1497620206.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "calendar_set", + "choice_b": "play_podcasts", + "choice_c": "hue_lightoff", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8166 + }, + { + "path": "audio--1504193120-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "play_game", + "choice_b": "alarm_query", + "choice_c": "general_greet", + "choice_d": "qa_maths", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8167 + }, + { + "path": "audio-1498571298.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "transport_taxi", + "choice_b": "joke", + "choice_c": "play_music", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8168 + }, + { + "path": "audio-1501752123.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "calendar_remove", + "choice_b": "datetime_convert", + "choice_c": "music_likeness", + "choice_d": "locations", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8169 + }, + { + "path": "audio-1501152304-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "qa_definition", + "choice_b": "music_settings", + "choice_c": "transport_query", + "choice_d": "audio_volume_down", + "answer_gt": "audio_volume_down", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8170 + }, + { + "path": "audio-1497025362-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "alarm_remove", + "choice_b": "play_podcasts", + "choice_c": "createoradd", + "choice_d": "email_addcontact", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8171 + }, + { + "path": "audio-1501756623.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "iot_hue_lighton", + "choice_b": "email_query", + "choice_c": "iot_wemo_off", + "choice_d": "play_radio", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8172 + }, + { + "path": "audio-1494512684-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "takeaway_order", + "choice_b": "weather_query", + "choice_c": "datetime_query", + "choice_d": "datetime_convert", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8173 + }, + { + "path": "audio-1500899973.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "play_audiobook", + "choice_b": "play_game", + "choice_c": "social_query", + "choice_d": "takeaway_order", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8174 + }, + { + "path": "audio-1497558430-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "calendar_set", + "choice_b": "iot_cleaning", + "choice_c": "joke", + "choice_d": "datetime_query", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8175 + }, + { + "path": "audio-1502195789.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "music", + "choice_b": "general_joke", + "choice_c": "calendar_remove", + "choice_d": "joke", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8176 + }, + { + "path": "audio-1492780957-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "qa_definition", + "choice_b": "email_query", + "choice_c": "calendar_set", + "choice_d": "play_radio", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8177 + }, + { + "path": "audio-1490282220-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "iot_coffee", + "choice_b": "createoradd", + "choice_c": "general_quirky", + "choice_d": "datetime_convert", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8178 + }, + { + "path": "audio-1500978354.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "lists_query", + "choice_b": "takeaway_query", + "choice_c": "email_addcontact", + "choice_d": "music_likeness", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8179 + }, + { + "path": "audio-1502101123.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "iot_coffee", + "choice_b": "play_podcasts", + "choice_c": "iot_hue_lightchange", + "choice_d": "recommendation_events", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8180 + }, + { + "path": "audio-1502300286-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "general_quirky", + "choice_b": "recommendation_events", + "choice_c": "locations", + "choice_d": "transport_query", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8181 + }, + { + "path": "audio-1498485487.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "weather_query", + "choice_b": "general_joke", + "choice_c": "general_quirky", + "choice_d": "play_music", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8182 + }, + { + "path": "audio-1502200378-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "general_quirky", + "choice_b": "general_greet", + "choice_c": "music_query", + "choice_d": "alarm_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8183 + }, + { + "path": "audio-1434528983-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "play_game", + "choice_b": "calendar_set", + "choice_c": "lists_remove", + "choice_d": "datetime_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8184 + }, + { + "path": "audio-1501687155-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "hue_lightoff", + "choice_b": "takeaway_query", + "choice_c": "music_likeness", + "choice_d": "createoradd", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8185 + }, + { + "path": "audio--1506079296-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "takeaway_query", + "choice_b": "qa_factoid", + "choice_c": "play_radio", + "choice_d": "weather_query", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8186 + }, + { + "path": "audio-1501697240-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "audio_volume_mute", + "choice_b": "calendar_query", + "choice_c": "qa_currency", + "choice_d": "iot_wemo_on", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8187 + }, + { + "path": "audio-1434535796.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "recommendation_events", + "choice_b": "transport_traffic", + "choice_c": "sendemail", + "choice_d": "social_post", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8188 + }, + { + "path": "audio-1502374982.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "qa_definition", + "choice_b": "transport_traffic", + "choice_c": "weather_query", + "choice_d": "cooking_recipe", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8189 + }, + { + "path": "audio--1504190258.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "play_radio", + "choice_b": "iot_coffee", + "choice_c": "calendar_set", + "choice_d": "datetime_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8190 + }, + { + "path": "audio-1494418111.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "recommendation_locations", + "choice_b": "general_joke", + "choice_c": "general_quirky", + "choice_d": "email_addcontact", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8191 + }, + { + "path": "audio-1490289532-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "music_settings", + "choice_b": "datetime_query", + "choice_c": "iot_coffee", + "choice_d": "audio_volume_down", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8192 + }, + { + "path": "audio-1501753448.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "iot_cleaning", + "choice_b": "news_query", + "choice_c": "iot_hue_lightup", + "choice_d": "alarm_set", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8193 + }, + { + "path": "audio-1434541340.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "transport_query", + "choice_b": "query", + "choice_c": "calendar_query", + "choice_d": "lists_remove", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8194 + }, + { + "path": "audio-1501758369.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "audio_volume_down", + "choice_b": "qa_definition", + "choice_c": "email_querycontact", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8195 + }, + { + "path": "audio-1494512621-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "music_settings", + "choice_b": "recommendation_locations", + "choice_c": "audio_volume_mute", + "choice_d": "iot_hue_lighton", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8196 + }, + { + "path": "audio-1490356169-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "iot_coffee", + "choice_b": "social_query", + "choice_c": "music_likeness", + "choice_d": "transport_query", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8197 + }, + { + "path": "audio-1501772111-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "post", + "choice_b": "qa_factoid", + "choice_c": "general_quirky", + "choice_d": "datetime_query", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8198 + }, + { + "path": "audio-1492757251.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "general_quirky", + "choice_b": "play_music", + "choice_c": "calendar_remove", + "choice_d": "iot_hue_lightup", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8199 + }, + { + "path": "audio-1497872401.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "general_quirky", + "choice_b": "iot_hue_lightdim", + "choice_c": "hue_lightdim", + "choice_d": "radio", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8200 + }, + { + "path": "audio-1497439935-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "qa_factoid", + "choice_b": "music_settings", + "choice_c": "audio_volume_down", + "choice_d": "joke", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8201 + }, + { + "path": "audio-1502195930-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "takeaway_query", + "choice_b": "joke", + "choice_c": "social_post", + "choice_d": "play_podcasts", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8202 + }, + { + "path": "audio-1501151467-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "general_quirky", + "choice_b": "play_podcasts", + "choice_c": "audio_volume_mute", + "choice_d": "play_music", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8203 + }, + { + "path": "audio-1434533567.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "radio", + "choice_b": "alarm_set", + "choice_c": "play_podcasts", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8204 + }, + { + "path": "audio-1499090503.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "locations", + "choice_b": "general_joke", + "choice_c": "music_likeness", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8205 + }, + { + "path": "audio-1497558241.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "calendar_set", + "choice_b": "iot_hue_lightoff", + "choice_c": "play_podcasts", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8206 + }, + { + "path": "audio-1488974854.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "sendemail", + "choice_c": "email_query", + "choice_d": "music", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8207 + }, + { + "path": "audio-1490262771-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "play_radio", + "choice_b": "transport_ticket", + "choice_c": "recommendation_locations", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8208 + }, + { + "path": "audio--1504193275.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "transport_query", + "choice_b": "locations", + "choice_c": "audio_volume_mute", + "choice_d": "qa_stock", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8209 + }, + { + "path": "audio-1501754205-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "qa_maths", + "choice_b": "datetime_convert", + "choice_c": "iot_coffee", + "choice_d": "audio_volume_up", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8210 + }, + { + "path": "audio--1504190572-headset.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "datetime_query", + "choice_b": "calendar_query", + "choice_c": "music_settings", + "choice_d": "play_radio", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8211 + }, + { + "path": "audio-1490106981.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "play_audiobook", + "choice_b": "play_game", + "choice_c": "general_greet", + "choice_d": "audio_volume_other", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8212 + }, + { + "path": "audio-1501777588-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "takeaway_order", + "choice_b": "lists_createoradd", + "choice_c": "lists_remove", + "choice_d": "qa_factoid", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8213 + }, + { + "path": "audio-1490800878.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "takeaway_order", + "choice_b": "qa_factoid", + "choice_c": "music", + "choice_d": "sendemail", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8214 + }, + { + "path": "audio-1501751693-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "alarm_set", + "choice_b": "recommendation_locations", + "choice_c": "transport_ticket", + "choice_d": "takeaway_query", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8215 + }, + { + "path": "audio--1504194734-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "music_likeness", + "choice_b": "play_game", + "choice_c": "qa_stock", + "choice_d": "alarm_remove", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8216 + }, + { + "path": "audio-1490182575-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "general_quirky", + "choice_b": "play_radio", + "choice_c": "calendar_set", + "choice_d": "alarm_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8217 + }, + { + "path": "audio-1490261246-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "iot_hue_lightoff", + "choice_b": "audio_volume_up", + "choice_c": "general_quirky", + "choice_d": "alarm_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8218 + }, + { + "path": "audio-1488969846.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "remove", + "choice_b": "hue_lightoff", + "choice_c": "calendar_set", + "choice_d": "play_game", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8219 + }, + { + "path": "audio-1499267108-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "transport_ticket", + "choice_b": "news_query", + "choice_c": "alarm_remove", + "choice_d": "transport_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8220 + }, + { + "path": "audio-1490182401.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "general_quirky", + "choice_b": "lists_createoradd", + "choice_c": "qa_stock", + "choice_d": "transport_taxi", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8221 + }, + { + "path": "audio-1502890356.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "music", + "choice_b": "general_joke", + "choice_c": "email_query", + "choice_d": "lists_remove", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8222 + }, + { + "path": "audio-1490183698-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "news_query", + "choice_b": "calendar_set", + "choice_c": "play_game", + "choice_d": "audio_volume_down", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8223 + }, + { + "path": "audio-1502199657.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "recommendation_movies", + "choice_b": "recommendation_locations", + "choice_c": "iot_hue_lightoff", + "choice_d": "transport_traffic", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8224 + }, + { + "path": "audio-1501156274.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "recommendation_events", + "choice_b": "datetime_query", + "choice_c": "qa_definition", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8225 + }, + { + "path": "audio--1506521190.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "recommendation_locations", + "choice_b": "audio_volume_other", + "choice_c": "alarm_query", + "choice_d": "music_likeness", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8226 + }, + { + "path": "audio-1490106506-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "lists_remove", + "choice_b": "iot_hue_lightchange", + "choice_c": "play_radio", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8227 + }, + { + "path": "audio-1499695891.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "transport_query", + "choice_b": "qa_maths", + "choice_c": "alarm_set", + "choice_d": "social_post", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8228 + }, + { + "path": "audio-1502200746-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "general_joke", + "choice_b": "query", + "choice_c": "hue_lightdim", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8229 + }, + { + "path": "audio-1498566650.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "radio", + "choice_b": "music_dislikeness", + "choice_c": "query", + "choice_d": "audio_volume_down", + "answer_gt": "music_dislikeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8230 + }, + { + "path": "audio-1501694763.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "news_query", + "choice_b": "calendar_set", + "choice_c": "audio_volume_other", + "choice_d": "cooking_recipe", + "answer_gt": "audio_volume_other", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8231 + }, + { + "path": "audio-1494421815-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "datetime_convert", + "choice_b": "play_radio", + "choice_c": "iot_hue_lightdim", + "choice_d": "iot_hue_lighton", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8232 + }, + { + "path": "audio-1502376487.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "transport_traffic", + "choice_b": "iot_wemo_off", + "choice_c": "qa_factoid", + "choice_d": "play_audiobook", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8233 + }, + { + "path": "audio-1501686601-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "play_music", + "choice_b": "transport_ticket", + "choice_c": "hue_lightoff", + "choice_d": "email_addcontact", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8234 + }, + { + "path": "audio-1497024957-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "qa_factoid", + "choice_b": "weather_query", + "choice_c": "email_addcontact", + "choice_d": "email_query", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8235 + }, + { + "path": "audio-1501688292-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "calendar_query", + "choice_b": "iot_wemo_on", + "choice_c": "qa_factoid", + "choice_d": "transport_traffic", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8236 + }, + { + "path": "audio-1490286956-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "play_podcasts", + "choice_b": "general_greet", + "choice_c": "iot_hue_lightoff", + "choice_d": "audio_volume_down", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8237 + }, + { + "path": "audio-1498567543-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "play_radio", + "choice_b": "iot_wemo_on", + "choice_c": "remove", + "choice_d": "transport_taxi", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8238 + }, + { + "path": "audio-1502376028.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "iot_hue_lighton", + "choice_b": "music_query", + "choice_c": "createoradd", + "choice_d": "audio_volume_other", + "answer_gt": "createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8239 + }, + { + "path": "audio-1498566094.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "news_query", + "choice_b": "music_likeness", + "choice_c": "play_music", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8240 + }, + { + "path": "audio-1495733243-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "alarm_query", + "choice_c": "general_quirky", + "choice_d": "iot_wemo_on", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8241 + }, + { + "path": "audio-1494416296.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "cooking_recipe", + "choice_b": "joke", + "choice_c": "remove", + "choice_d": "music_dislikeness", + "answer_gt": "joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8242 + }, + { + "path": "audio-1498572193.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "email_addcontact", + "choice_b": "iot_hue_lightchange", + "choice_c": "iot_coffee", + "choice_d": "qa_stock", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8243 + }, + { + "path": "audio-1502375323-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "email_query", + "choice_c": "recommendation_locations", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8244 + }, + { + "path": "audio-1492784612.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "alarm_set", + "choice_b": "qa_definition", + "choice_c": "calendar_set", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8245 + }, + { + "path": "audio-1490370978.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "sendemail", + "choice_b": "social_post", + "choice_c": "email_sendemail", + "choice_d": "iot_coffee", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8246 + }, + { + "path": "audio-1490955378.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "audio_volume_down", + "choice_b": "play_game", + "choice_c": "qa_factoid", + "choice_d": "lists_query", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8247 + }, + { + "path": "audio-1489153386.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "transport_query", + "choice_b": "iot_wemo_off", + "choice_c": "play_music", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8248 + }, + { + "path": "audio-1488972176.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "recommendation_events", + "choice_b": "cooking_recipe", + "choice_c": "iot_wemo_off", + "choice_d": "audio_volume_mute", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8249 + }, + { + "path": "audio-1488983412.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "iot_hue_lighton", + "choice_b": "qa_stock", + "choice_c": "iot_hue_lightdim", + "choice_d": "calendar_set", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8250 + }, + { + "path": "audio-1501777748.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "iot_hue_lighton", + "choice_c": "social_post", + "choice_d": "alarm_set", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8251 + }, + { + "path": "audio-1490261437.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "qa_maths", + "choice_b": "iot_hue_lightoff", + "choice_c": "audio_volume_down", + "choice_d": "email_query", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8252 + }, + { + "path": "audio-1501775574.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "joke", + "choice_b": "sendemail", + "choice_c": "iot_wemo_on", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8253 + }, + { + "path": "audio-1500036829-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "play_podcasts", + "choice_b": "audio_volume_mute", + "choice_c": "datetime_query", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8254 + }, + { + "path": "audio-1501156668.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "recommendation_movies", + "choice_b": "alarm_set", + "choice_c": "createoradd", + "choice_d": "post", + "answer_gt": "post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8255 + }, + { + "path": "audio-1490287459.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "lists_createoradd", + "choice_b": "play_podcasts", + "choice_c": "recommendation_movies", + "choice_d": "iot_hue_lighton", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8256 + }, + { + "path": "audio-1499692133.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "takeaway_order", + "choice_b": "qa_factoid", + "choice_c": "recommendation_events", + "choice_d": "qa_stock", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8257 + }, + { + "path": "audio-1490967852-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "play_audiobook", + "choice_b": "calendar_remove", + "choice_c": "takeaway_order", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8258 + }, + { + "path": "audio-1434534525-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "email_addcontact", + "choice_b": "transport_taxi", + "choice_c": "iot_wemo_on", + "choice_d": "qa_currency", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8259 + }, + { + "path": "audio-1497880762-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "iot_hue_lighton", + "choice_b": "transport_taxi", + "choice_c": "calendar_remove", + "choice_d": "audio_volume_up", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8260 + }, + { + "path": "audio-1488984407.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_hue_lighton", + "choice_b": "takeaway_query", + "choice_c": "qa_currency", + "choice_d": "hue_lightoff", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8261 + }, + { + "path": "audio-1434540486.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "remove", + "choice_b": "email_querycontact", + "choice_c": "datetime_convert", + "choice_d": "general_joke", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8262 + }, + { + "path": "audio-1490200791.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "audio_volume_down", + "choice_b": "iot_coffee", + "choice_c": "qa_factoid", + "choice_d": "joke", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8263 + }, + { + "path": "audio-1502895066.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_stock", + "choice_b": "joke", + "choice_c": "iot_hue_lightchange", + "choice_d": "play_audiobook", + "answer_gt": "qa_stock", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8264 + }, + { + "path": "audio-1498573594.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "datetime_query", + "choice_b": "social_query", + "choice_c": "music_likeness", + "choice_d": "hue_lightdim", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8265 + }, + { + "path": "audio-1502101733.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "email_query", + "choice_c": "general_joke", + "choice_d": "qa_currency", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8266 + }, + { + "path": "audio-1497622747-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "joke", + "choice_b": "radio", + "choice_c": "alarm_set", + "choice_d": "play_podcasts", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8267 + }, + { + "path": "audio-1497879813.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "qa_currency", + "choice_b": "general_greet", + "choice_c": "datetime_convert", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8268 + }, + { + "path": "audio-1490195512-headset.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "query", + "choice_b": "play_radio", + "choice_c": "lists_query", + "choice_d": "iot_coffee", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8269 + }, + { + "path": "audio-1490193389-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "lists_query", + "choice_b": "play_podcasts", + "choice_c": "general_quirky", + "choice_d": "hue_lightdim", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8270 + }, + { + "path": "audio-1494418766-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "query", + "choice_b": "alarm_set", + "choice_c": "calendar_set", + "choice_d": "general_greet", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8271 + }, + { + "path": "audio-1490357294-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "query", + "choice_b": "email_query", + "choice_c": "transport_query", + "choice_d": "music", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8272 + }, + { + "path": "audio--1504192211.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "calendar_query", + "choice_b": "play_audiobook", + "choice_c": "qa_factoid", + "choice_d": "email_querycontact", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8273 + }, + { + "path": "audio-1502101442-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "lists_createoradd", + "choice_b": "music", + "choice_c": "iot_hue_lightchange", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8274 + }, + { + "path": "audio-1501754224-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "qa_factoid", + "choice_b": "iot_hue_lightchange", + "choice_c": "qa_definition", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8275 + }, + { + "path": "audio-1501775130.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "iot_hue_lightchange", + "choice_b": "datetime_convert", + "choice_c": "general_quirky", + "choice_d": "audio_volume_down", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8276 + }, + { + "path": "audio-1490105403-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "transport_taxi", + "choice_b": "play_audiobook", + "choice_c": "sendemail", + "choice_d": "iot_cleaning", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8277 + }, + { + "path": "audio--1504192184-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "general_joke", + "choice_b": "alarm_set", + "choice_c": "calendar_set", + "choice_d": "general_greet", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8278 + }, + { + "path": "audio-1499688437.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "lists_createoradd", + "choice_b": "music", + "choice_c": "news_query", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8279 + }, + { + "path": "audio-1488993308.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "calendar_set", + "choice_c": "general_quirky", + "choice_d": "transport_taxi", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8280 + }, + { + "path": "audio-1502375072-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "iot_coffee", + "choice_b": "email_sendemail", + "choice_c": "audio_volume_down", + "choice_d": "email_query", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8281 + }, + { + "path": "audio-1499089782.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "music_query", + "choice_b": "news_query", + "choice_c": "sendemail", + "choice_d": "qa_factoid", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8282 + }, + { + "path": "audio-1498573267-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "music_settings", + "choice_b": "qa_definition", + "choice_c": "general_joke", + "choice_d": "play_music", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8283 + }, + { + "path": "audio-1434539841.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "calendar_set", + "choice_b": "datetime_convert", + "choice_c": "radio", + "choice_d": "weather_query", + "answer_gt": "datetime_convert", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8284 + }, + { + "path": "audio-1501777268.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "takeaway_query", + "choice_b": "qa_stock", + "choice_c": "calendar_set", + "choice_d": "transport_ticket", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8285 + }, + { + "path": "audio-1490183640.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "weather_query", + "choice_b": "calendar_remove", + "choice_c": "iot_hue_lightdim", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8286 + }, + { + "path": "audio-1502299633.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "locations", + "choice_b": "weather_query", + "choice_c": "qa_maths", + "choice_d": "audio_volume_other", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8287 + }, + { + "path": "audio-1498757508.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_hue_lighton", + "choice_b": "music", + "choice_c": "email_query", + "choice_d": "alarm_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8288 + }, + { + "path": "audio-1501764223.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "qa_maths", + "choice_b": "recommendation_locations", + "choice_c": "email_query", + "choice_d": "general_joke", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8289 + }, + { + "path": "audio-1502198108-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "news_query", + "choice_b": "audio_volume_other", + "choice_c": "play_radio", + "choice_d": "joke", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8290 + }, + { + "path": "audio-1502100588.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "social_post", + "choice_b": "social_query", + "choice_c": "audio_volume_up", + "choice_d": "iot_hue_lightchange", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8291 + }, + { + "path": "audio-1495371341-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_hue_lighton", + "choice_b": "qa_maths", + "choice_c": "datetime_query", + "choice_d": "iot_hue_lightdim", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8292 + }, + { + "path": "audio-1492785833-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "qa_factoid", + "choice_b": "audio_volume_mute", + "choice_c": "music_settings", + "choice_d": "iot_hue_lightoff", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8293 + }, + { + "path": "audio-1502101733-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "transport_taxi", + "choice_b": "qa_factoid", + "choice_c": "qa_currency", + "choice_d": "music", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8294 + }, + { + "path": "audio-1501414539.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "audio_volume_other", + "choice_b": "datetime_query", + "choice_c": "qa_definition", + "choice_d": "social_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8295 + }, + { + "path": "audio-1497880339.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "hue_lightdim", + "choice_b": "recommendation_events", + "choice_c": "play_music", + "choice_d": "qa_currency", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8296 + }, + { + "path": "audio-1502896975-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "joke", + "choice_c": "play_game", + "choice_d": "takeaway_order", + "answer_gt": "iot_hue_lightup", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8297 + }, + { + "path": "audio-1501694653-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "weather_query", + "choice_b": "iot_hue_lightdim", + "choice_c": "lists_createoradd", + "choice_d": "music_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8298 + }, + { + "path": "audio-1494607541-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "hue_lightoff", + "choice_b": "social_query", + "choice_c": "iot_hue_lightchange", + "choice_d": "hue_lightdim", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8299 + }, + { + "path": "audio-1488974371.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "email_addcontact", + "choice_b": "iot_hue_lightoff", + "choice_c": "audio_volume_up", + "choice_d": "audio_volume_mute", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8300 + }, + { + "path": "audio--1505406801.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "play_audiobook", + "choice_b": "qa_stock", + "choice_c": "transport_taxi", + "choice_d": "audio_volume_down", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8301 + }, + { + "path": "audio-1502101373-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "general_quirky", + "choice_b": "qa_currency", + "choice_c": "locations", + "choice_d": "iot_wemo_on", + "answer_gt": "iot_wemo_on", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8302 + }, + { + "path": "audio-1490109166-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "query", + "choice_b": "play_podcasts", + "choice_c": "music", + "choice_d": "general_joke", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8303 + }, + { + "path": "audio-1500387110-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "alarm_query", + "choice_b": "qa_definition", + "choice_c": "play_radio", + "choice_d": "createoradd", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8304 + }, + { + "path": "audio-1501754644.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "audio_volume_up", + "choice_b": "calendar_set", + "choice_c": "sendemail", + "choice_d": "iot_hue_lightup", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8305 + }, + { + "path": "audio-1490193091-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "weather_query", + "choice_b": "play_audiobook", + "choice_c": "iot_wemo_on", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8306 + }, + { + "path": "audio-1490283449-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "transport_traffic", + "choice_b": "music_dislikeness", + "choice_c": "qa_stock", + "choice_d": "social_post", + "answer_gt": "qa_stock", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8307 + }, + { + "path": "audio--1504195260-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "music_likeness", + "choice_b": "social_post", + "choice_c": "lists_query", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8308 + }, + { + "path": "audio-1502376962.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "transport_ticket", + "choice_b": "hue_lightoff", + "choice_c": "general_joke", + "choice_d": "play_music", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8309 + }, + { + "path": "audio-1490956441.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "locations", + "choice_b": "recommendation_events", + "choice_c": "news_query", + "choice_d": "calendar_remove", + "answer_gt": "locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8310 + }, + { + "path": "audio-1499246103-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "social_post", + "choice_b": "audio_volume_other", + "choice_c": "iot_hue_lightchange", + "choice_d": "qa_factoid", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8311 + }, + { + "path": "audio-1497620862-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "music_likeness", + "choice_b": "iot_hue_lightchange", + "choice_c": "recommendation_events", + "choice_d": "transport_query", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8312 + }, + { + "path": "audio-1502196125-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "iot_hue_lightoff", + "choice_b": "alarm_remove", + "choice_c": "weather_query", + "choice_d": "email_querycontact", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8313 + }, + { + "path": "audio-1488562784.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "play_game", + "choice_b": "iot_hue_lightchange", + "choice_c": "alarm_set", + "choice_d": "music_settings", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8314 + }, + { + "path": "audio-1490801069-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "iot_hue_lightdim", + "choice_b": "recommendation_locations", + "choice_c": "createoradd", + "choice_d": "hue_lightoff", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8315 + }, + { + "path": "audio-1488993668.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "qa_maths", + "choice_b": "general_joke", + "choice_c": "radio", + "choice_d": "qa_definition", + "answer_gt": "radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8316 + }, + { + "path": "audio-1490194380-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "qa_stock", + "choice_b": "alarm_remove", + "choice_c": "sendemail", + "choice_d": "recommendation_movies", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8317 + }, + { + "path": "audio-1502100337.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "audio_volume_down", + "choice_b": "cooking_recipe", + "choice_c": "social_post", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8318 + }, + { + "path": "audio-1490356700-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "calendar_set", + "choice_b": "qa_factoid", + "choice_c": "transport_traffic", + "choice_d": "lists_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8319 + }, + { + "path": "audio-1498572946.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "transport_traffic", + "choice_b": "takeaway_order", + "choice_c": "lists_remove", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8320 + }, + { + "path": "audio--1506078335.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "play_podcasts", + "choice_b": "transport_ticket", + "choice_c": "lists_query", + "choice_d": "play_game", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8321 + }, + { + "path": "audio-1488552442.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "audio_volume_other", + "choice_b": "iot_hue_lighton", + "choice_c": "weather_query", + "choice_d": "calendar_set", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8322 + }, + { + "path": "audio-1501756945.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "iot_coffee", + "choice_b": "query", + "choice_c": "iot_hue_lightup", + "choice_d": "recommendation_locations", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8323 + }, + { + "path": "audio-1495376966.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "iot_hue_lightup", + "choice_b": "news_query", + "choice_c": "datetime_convert", + "choice_d": "iot_wemo_off", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8324 + }, + { + "path": "audio-1501691005.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "qa_factoid", + "choice_b": "datetime_query", + "choice_c": "play_podcasts", + "choice_d": "datetime_convert", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8325 + }, + { + "path": "audio-1490283755.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "qa_factoid", + "choice_b": "joke", + "choice_c": "calendar_set", + "choice_d": "music_dislikeness", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8326 + }, + { + "path": "audio--1506078485.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "music", + "choice_b": "qa_maths", + "choice_c": "alarm_set", + "choice_d": "social_query", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8327 + }, + { + "path": "audio-1497622747.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "recommendation_events", + "choice_b": "iot_hue_lightup", + "choice_c": "audio_volume_up", + "choice_d": "alarm_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8328 + }, + { + "path": "audio-1497862425-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "iot_hue_lightchange", + "choice_b": "datetime_query", + "choice_c": "audio_volume_other", + "choice_d": "play_audiobook", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8329 + }, + { + "path": "audio-1501698525-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "social_post", + "choice_b": "play_music", + "choice_c": "audio_volume_up", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8330 + }, + { + "path": "audio-1490202240-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "lists_createoradd", + "choice_b": "alarm_query", + "choice_c": "play_game", + "choice_d": "social_post", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8331 + }, + { + "path": "audio-1494512860.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "play_music", + "choice_b": "social_query", + "choice_c": "qa_definition", + "choice_d": "qa_stock", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8332 + }, + { + "path": "audio-1490194916.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "qa_factoid", + "choice_b": "social_post", + "choice_c": "calendar_set", + "choice_d": "createoradd", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8333 + }, + { + "path": "audio-1490193477-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "play_audiobook", + "choice_b": "audio_volume_mute", + "choice_c": "qa_factoid", + "choice_d": "lists_query", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8334 + }, + { + "path": "audio-1490104593.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "lists_query", + "choice_b": "lists_remove", + "choice_c": "email_sendemail", + "choice_d": "remove", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8335 + }, + { + "path": "audio-1497360787.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "datetime_query", + "choice_b": "play_music", + "choice_c": "email_querycontact", + "choice_d": "play_podcasts", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8336 + }, + { + "path": "audio-1494549077.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_currency", + "choice_b": "iot_hue_lighton", + "choice_c": "qa_definition", + "choice_d": "play_radio", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8337 + }, + { + "path": "audio-1494608905.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "iot_coffee", + "choice_b": "qa_factoid", + "choice_c": "calendar_set", + "choice_d": "music_dislikeness", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8338 + }, + { + "path": "audio-1501758871-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "qa_definition", + "choice_b": "transport_ticket", + "choice_c": "music_likeness", + "choice_d": "play_audiobook", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8339 + }, + { + "path": "audio-1490104820-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "email_querycontact", + "choice_b": "cooking_recipe", + "choice_c": "iot_hue_lighton", + "choice_d": "weather_query", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8340 + }, + { + "path": "audio-1490103669.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "transport_taxi", + "choice_b": "takeaway_order", + "choice_c": "recommendation_locations", + "choice_d": "post", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8341 + }, + { + "path": "audio-1498572917.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "hue_lightdim", + "choice_b": "general_greet", + "choice_c": "iot_wemo_off", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8342 + }, + { + "path": "audio-1490954650.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "general_joke", + "choice_b": "music", + "choice_c": "email_sendemail", + "choice_d": "calendar_remove", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8343 + }, + { + "path": "audio-1497883484.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "sendemail", + "choice_b": "play_audiobook", + "choice_c": "transport_taxi", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8344 + }, + { + "path": "audio-1490107043.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "iot_wemo_on", + "choice_b": "email_sendemail", + "choice_c": "transport_traffic", + "choice_d": "transport_ticket", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8345 + }, + { + "path": "audio-1502195485.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "iot_wemo_on", + "choice_b": "play_podcasts", + "choice_c": "music_likeness", + "choice_d": "email_addcontact", + "answer_gt": "iot_wemo_on", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8346 + }, + { + "path": "audio-1490263665-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "iot_wemo_on", + "choice_c": "remove", + "choice_d": "alarm_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8347 + }, + { + "path": "audio-1490183575.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "recommendation_movies", + "choice_b": "social_query", + "choice_c": "iot_coffee", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8348 + }, + { + "path": "audio-1490182883.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "query", + "choice_b": "news_query", + "choice_c": "email_querycontact", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8349 + }, + { + "path": "audio-1497860844-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "audio_volume_down", + "choice_b": "music_dislikeness", + "choice_c": "query", + "choice_d": "iot_coffee", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8350 + }, + { + "path": "audio-1490111030-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "joke", + "choice_b": "lists_query", + "choice_c": "transport_traffic", + "choice_d": "post", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8351 + }, + { + "path": "audio-1497030882.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "music_query", + "choice_b": "calendar_query", + "choice_c": "general_quirky", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8352 + }, + { + "path": "audio-1502896210.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "takeaway_query", + "choice_b": "email_query", + "choice_c": "recommendation_locations", + "choice_d": "news_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8353 + }, + { + "path": "audio-1501698732.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "audio_volume_other", + "choice_b": "recommendation_locations", + "choice_c": "audio_volume_mute", + "choice_d": "qa_maths", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8354 + }, + { + "path": "audio-1490104821-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "transport_ticket", + "choice_b": "calendar_set", + "choice_c": "general_quirky", + "choice_d": "recommendation_locations", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8355 + }, + { + "path": "audio-1488542627.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "iot_hue_lightchange", + "choice_b": "iot_hue_lighton", + "choice_c": "qa_maths", + "choice_d": "alarm_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8356 + }, + { + "path": "audio-1498572758.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "general_greet", + "choice_b": "recommendation_events", + "choice_c": "weather_query", + "choice_d": "social_post", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8357 + }, + { + "path": "audio-1434531780-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "play_audiobook", + "choice_b": "query", + "choice_c": "play_radio", + "choice_d": "radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8358 + }, + { + "path": "audio-1502101334-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "alarm_remove", + "choice_c": "takeaway_query", + "choice_d": "datetime_query", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8359 + }, + { + "path": "audio--1504197546-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "iot_wemo_on", + "choice_b": "audio_volume_up", + "choice_c": "alarm_query", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8360 + }, + { + "path": "audio-1490202859-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "recommendation_locations", + "choice_b": "lists_createoradd", + "choice_c": "play_music", + "choice_d": "transport_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8361 + }, + { + "path": "audio-1502300219-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "calendar_set", + "choice_b": "radio", + "choice_c": "transport_traffic", + "choice_d": "general_greet", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8362 + }, + { + "path": "audio--1504195339-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "qa_factoid", + "choice_b": "transport_ticket", + "choice_c": "social_post", + "choice_d": "audio_volume_down", + "answer_gt": "audio_volume_down", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8363 + }, + { + "path": "audio-1501758710.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "hue_lightdim", + "choice_b": "iot_hue_lightchange", + "choice_c": "play_radio", + "choice_d": "calendar_set", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8364 + }, + { + "path": "audio-1490289514.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "calendar_remove", + "choice_b": "recommendation_movies", + "choice_c": "calendar_set", + "choice_d": "transport_ticket", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8365 + }, + { + "path": "audio-1501772231.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "email_addcontact", + "choice_b": "general_quirky", + "choice_c": "weather_query", + "choice_d": "transport_taxi", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8366 + }, + { + "path": "audio-1490106525.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "music_dislikeness", + "choice_b": "iot_hue_lightdim", + "choice_c": "calendar_set", + "choice_d": "email_querycontact", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8367 + }, + { + "path": "audio-1501764034-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "music_query", + "choice_b": "play_game", + "choice_c": "transport_ticket", + "choice_d": "email_addcontact", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8368 + }, + { + "path": "audio-1488557062.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "takeaway_query", + "choice_b": "email_query", + "choice_c": "music_query", + "choice_d": "general_joke", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8369 + }, + { + "path": "audio-1434528649-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "social_post", + "choice_b": "alarm_set", + "choice_c": "audio_volume_up", + "choice_d": "alarm_query", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8370 + }, + { + "path": "audio-1502195035.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "lists_createoradd", + "choice_c": "iot_coffee", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8371 + }, + { + "path": "audio-1498574020.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "qa_currency", + "choice_b": "general_joke", + "choice_c": "locations", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8372 + }, + { + "path": "audio-1499697062-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "social_query", + "choice_c": "takeaway_order", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8373 + }, + { + "path": "audio-1502891193.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "qa_stock", + "choice_b": "qa_currency", + "choice_c": "audio_volume_mute", + "choice_d": "iot_wemo_off", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8374 + }, + { + "path": "audio-1501151339-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "lists_query", + "choice_b": "news_query", + "choice_c": "play_radio", + "choice_d": "datetime_convert", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8375 + }, + { + "path": "audio-1490195818-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "general_greet", + "choice_b": "qa_currency", + "choice_c": "iot_hue_lightchange", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8376 + }, + { + "path": "audio-1501693693.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "joke", + "choice_b": "hue_lightdim", + "choice_c": "email_addcontact", + "choice_d": "social_post", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8377 + }, + { + "path": "audio-1488983583.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "qa_definition", + "choice_b": "email_sendemail", + "choice_c": "qa_stock", + "choice_d": "transport_query", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8378 + }, + { + "path": "audio-1498574367-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "post", + "choice_b": "calendar_remove", + "choice_c": "iot_cleaning", + "choice_d": "query", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8379 + }, + { + "path": "audio-1434528846-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "lists_remove", + "choice_b": "general_joke", + "choice_c": "datetime_convert", + "choice_d": "iot_hue_lightchange", + "answer_gt": "datetime_convert", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8380 + }, + { + "path": "audio-1494418111-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "general_quirky", + "choice_b": "transport_query", + "choice_c": "qa_definition", + "choice_d": "email_addcontact", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8381 + }, + { + "path": "audio-1502892763.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "general_greet", + "choice_b": "general_quirky", + "choice_c": "play_audiobook", + "choice_d": "alarm_set", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8382 + }, + { + "path": "audio-1501775568.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "locations", + "choice_b": "email_addcontact", + "choice_c": "qa_currency", + "choice_d": "iot_wemo_off", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8383 + }, + { + "path": "audio--1504192293-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "audio_volume_up", + "choice_b": "takeaway_query", + "choice_c": "transport_ticket", + "choice_d": "iot_wemo_on", + "answer_gt": "audio_volume_up", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8384 + }, + { + "path": "audio--1506521975.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "sendemail", + "choice_c": "audio_volume_mute", + "choice_d": "general_greet", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8385 + }, + { + "path": "audio-1490968061-headset.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "calendar_set", + "choice_c": "recommendation_movies", + "choice_d": "takeaway_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8386 + }, + { + "path": "audio-1501778782-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "music_likeness", + "choice_b": "qa_currency", + "choice_c": "social_post", + "choice_d": "sendemail", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8387 + }, + { + "path": "audio-1501155969.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "play_audiobook", + "choice_b": "lists_createoradd", + "choice_c": "weather_query", + "choice_d": "iot_hue_lightoff", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8388 + }, + { + "path": "audio-1494548568.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "joke", + "choice_b": "qa_definition", + "choice_c": "recommendation_movies", + "choice_d": "sendemail", + "answer_gt": "sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8389 + }, + { + "path": "audio-1498565714-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "music", + "choice_b": "recommendation_locations", + "choice_c": "general_quirky", + "choice_d": "lists_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8390 + }, + { + "path": "audio-1488973982.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_wemo_on", + "choice_b": "calendar_remove", + "choice_c": "lists_createoradd", + "choice_d": "qa_maths", + "answer_gt": "iot_wemo_on", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8391 + }, + { + "path": "audio-1498574020-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "news_query", + "choice_b": "audio_volume_mute", + "choice_c": "calendar_set", + "choice_d": "play_radio", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8392 + }, + { + "path": "audio-1490369278.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "createoradd", + "choice_b": "calendar_remove", + "choice_c": "recommendation_locations", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8393 + }, + { + "path": "audio-1501763753.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "sendemail", + "choice_b": "social_query", + "choice_c": "hue_lightoff", + "choice_d": "post", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8394 + }, + { + "path": "audio-1501688896-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "music_dislikeness", + "choice_b": "news_query", + "choice_c": "alarm_remove", + "choice_d": "remove", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8395 + }, + { + "path": "audio-1490007571-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "music_dislikeness", + "choice_b": "qa_currency", + "choice_c": "general_greet", + "choice_d": "play_podcasts", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8396 + }, + { + "path": "audio-1502376685.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "iot_wemo_on", + "choice_b": "calendar_query", + "choice_c": "transport_traffic", + "choice_d": "createoradd", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8397 + }, + { + "path": "audio-1501691687-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "post", + "choice_b": "hue_lightoff", + "choice_c": "calendar_set", + "choice_d": "calendar_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8398 + }, + { + "path": "audio-1497880173-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "calendar_set", + "choice_c": "weather_query", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8399 + }, + { + "path": "audio-1501692426-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "joke", + "choice_c": "music_settings", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8400 + }, + { + "path": "audio-1500039778.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "qa_maths", + "choice_b": "calendar_query", + "choice_c": "qa_currency", + "choice_d": "alarm_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8401 + }, + { + "path": "audio-1501695031-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "qa_stock", + "choice_b": "locations", + "choice_c": "calendar_remove", + "choice_d": "email_addcontact", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8402 + }, + { + "path": "audio-1497030693-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "recommendation_locations", + "choice_b": "email_query", + "choice_c": "sendemail", + "choice_d": "takeaway_order", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8403 + }, + { + "path": "audio-1500978527.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "createoradd", + "choice_b": "recommendation_locations", + "choice_c": "sendemail", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8404 + }, + { + "path": "audio-1494512655-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "music_dislikeness", + "choice_b": "remove", + "choice_c": "general_greet", + "choice_d": "play_podcasts", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8405 + }, + { + "path": "audio-1500979196.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "qa_currency", + "choice_b": "recommendation_locations", + "choice_c": "iot_hue_lightchange", + "choice_d": "qa_maths", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8406 + }, + { + "path": "audio-1502101313.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "transport_query", + "choice_b": "email_addcontact", + "choice_c": "iot_hue_lightdim", + "choice_d": "social_post", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8407 + }, + { + "path": "audio-1501698719.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_maths", + "choice_b": "hue_lightoff", + "choice_c": "iot_cleaning", + "choice_d": "music_likeness", + "answer_gt": "hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8408 + }, + { + "path": "audio-1502375014-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "email_sendemail", + "choice_b": "play_radio", + "choice_c": "play_podcasts", + "choice_d": "createoradd", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8409 + }, + { + "path": "audio-1501155175-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "email_sendemail", + "choice_b": "general_joke", + "choice_c": "recommendation_locations", + "choice_d": "weather_query", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8410 + }, + { + "path": "audio-1492780710-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "music_settings", + "choice_b": "datetime_query", + "choice_c": "audio_volume_other", + "choice_d": "qa_definition", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8411 + }, + { + "path": "audio-1489152266.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "iot_hue_lightup", + "choice_b": "iot_hue_lighton", + "choice_c": "alarm_remove", + "choice_d": "play_podcasts", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8412 + }, + { + "path": "audio-1492784810.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "music_dislikeness", + "choice_b": "email_query", + "choice_c": "datetime_convert", + "choice_d": "joke", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8413 + }, + { + "path": "audio-1501778757.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "play_audiobook", + "choice_b": "lists_remove", + "choice_c": "music_likeness", + "choice_d": "iot_hue_lighton", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8414 + }, + { + "path": "audio-1498574329.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "takeaway_order", + "choice_b": "general_joke", + "choice_c": "remove", + "choice_d": "cooking_recipe", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8415 + }, + { + "path": "audio-1490006977-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "music_query", + "choice_b": "qa_definition", + "choice_c": "play_radio", + "choice_d": "post", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8416 + }, + { + "path": "audio-1497867835-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "calendar_set", + "choice_b": "qa_factoid", + "choice_c": "weather_query", + "choice_d": "transport_traffic", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8417 + }, + { + "path": "audio-1501759249.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "datetime_convert", + "choice_b": "iot_cleaning", + "choice_c": "audio_volume_mute", + "choice_d": "calendar_set", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8418 + }, + { + "path": "audio-1488986652.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "post", + "choice_b": "play_audiobook", + "choice_c": "music_settings", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8419 + }, + { + "path": "audio-1434535796-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "iot_hue_lighton", + "choice_b": "recommendation_events", + "choice_c": "recommendation_movies", + "choice_d": "music_dislikeness", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8420 + }, + { + "path": "audio-1499244181-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "calendar_query", + "choice_b": "hue_lightoff", + "choice_c": "music_query", + "choice_d": "qa_currency", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8421 + }, + { + "path": "audio-1490371445-headset.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "music_dislikeness", + "choice_b": "iot_hue_lightchange", + "choice_c": "recommendation_movies", + "choice_d": "audio_volume_down", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8422 + }, + { + "path": "audio-1434542735.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "general_joke", + "choice_b": "news_query", + "choice_c": "weather_query", + "choice_d": "recommendation_locations", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8423 + }, + { + "path": "audio--1504195083-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "play_audiobook", + "choice_b": "music_likeness", + "choice_c": "alarm_set", + "choice_d": "iot_coffee", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8424 + }, + { + "path": "audio-1490370170-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "audio_volume_down", + "choice_b": "recommendation_locations", + "choice_c": "music_settings", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8425 + }, + { + "path": "audio-1498572402-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "hue_lightdim", + "choice_b": "iot_coffee", + "choice_c": "remove", + "choice_d": "cooking_recipe", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8426 + }, + { + "path": "audio-1499697011-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "music_likeness", + "choice_b": "iot_cleaning", + "choice_c": "music_dislikeness", + "choice_d": "calendar_query", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8427 + }, + { + "path": "audio-1498485748.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "transport_query", + "choice_b": "play_music", + "choice_c": "calendar_set", + "choice_d": "lists_remove", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8428 + }, + { + "path": "audio-1497865315-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "play_music", + "choice_b": "audio_volume_down", + "choice_c": "calendar_set", + "choice_d": "alarm_set", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8429 + }, + { + "path": "audio-1501752379-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "recommendation_events", + "choice_b": "datetime_query", + "choice_c": "music", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8430 + }, + { + "path": "audio-1492785405.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "audio_volume_down", + "choice_b": "iot_hue_lightchange", + "choice_c": "lists_query", + "choice_d": "sendemail", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8431 + }, + { + "path": "audio-1501768588-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "qa_maths", + "choice_b": "lists_createoradd", + "choice_c": "play_game", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8432 + }, + { + "path": "audio-1502302465.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "music_dislikeness", + "choice_b": "music", + "choice_c": "general_joke", + "choice_d": "recommendation_movies", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8433 + }, + { + "path": "audio-1495370327-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "play_music", + "choice_b": "audio_volume_down", + "choice_c": "audio_volume_up", + "choice_d": "calendar_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8434 + }, + { + "path": "audio-1490370692-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "play_audiobook", + "choice_b": "iot_wemo_off", + "choice_c": "music", + "choice_d": "play_podcasts", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8435 + }, + { + "path": "audio-1500036964.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "takeaway_query", + "choice_b": "music_settings", + "choice_c": "lists_createoradd", + "choice_d": "qa_maths", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8436 + }, + { + "path": "audio-1501697509.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "remove", + "choice_b": "qa_maths", + "choice_c": "general_greet", + "choice_d": "transport_ticket", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8437 + }, + { + "path": "audio-1497439388-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "email_addcontact", + "choice_b": "lists_remove", + "choice_c": "iot_wemo_on", + "choice_d": "play_radio", + "answer_gt": "iot_wemo_on", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8438 + }, + { + "path": "audio-1497436847-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "social_query", + "choice_b": "email_query", + "choice_c": "weather_query", + "choice_d": "takeaway_order", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8439 + }, + { + "path": "audio-1489499064.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "calendar_query", + "choice_b": "takeaway_query", + "choice_c": "music_likeness", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8440 + }, + { + "path": "audio-1490261089.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "iot_coffee", + "choice_b": "weather_query", + "choice_c": "calendar_set", + "choice_d": "general_joke", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8441 + }, + { + "path": "audio-1501768147-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "locations", + "choice_b": "iot_hue_lightdim", + "choice_c": "transport_traffic", + "choice_d": "qa_factoid", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8442 + }, + { + "path": "audio-1501158604-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "audio_volume_mute", + "choice_b": "music", + "choice_c": "audio_volume_down", + "choice_d": "play_podcasts", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8443 + }, + { + "path": "audio-1501686982.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "iot_wemo_off", + "choice_b": "general_quirky", + "choice_c": "music", + "choice_d": "qa_factoid", + "answer_gt": "music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8444 + }, + { + "path": "audio-1490110907.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "email_query", + "choice_b": "email_querycontact", + "choice_c": "alarm_remove", + "choice_d": "news_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8445 + }, + { + "path": "audio-1490359984-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "iot_wemo_on", + "choice_b": "email_sendemail", + "choice_c": "play_game", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8446 + }, + { + "path": "audio-1495377655-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "createoradd", + "choice_b": "iot_hue_lightchange", + "choice_c": "alarm_set", + "choice_d": "hue_lightdim", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8447 + }, + { + "path": "audio-1501757448-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "transport_taxi", + "choice_b": "weather_query", + "choice_c": "play_game", + "choice_d": "cooking_recipe", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8448 + }, + { + "path": "audio-1494415625-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "calendar_query", + "choice_b": "transport_query", + "choice_c": "qa_maths", + "choice_d": "iot_hue_lightup", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8449 + }, + { + "path": "audio-1502201394.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "lists_query", + "choice_b": "news_query", + "choice_c": "email_sendemail", + "choice_d": "play_music", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8450 + }, + { + "path": "audio--1504190646-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "cooking_recipe", + "choice_b": "transport_traffic", + "choice_c": "play_podcasts", + "choice_d": "iot_wemo_off", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8451 + }, + { + "path": "audio-1501608650.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "play_podcasts", + "choice_b": "weather_query", + "choice_c": "radio", + "choice_d": "cooking_recipe", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8452 + }, + { + "path": "audio-1501157563.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "news_query", + "choice_b": "qa_currency", + "choice_c": "recommendation_locations", + "choice_d": "play_music", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8453 + }, + { + "path": "audio-1501691252-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "hue_lightdim", + "choice_b": "hue_lightoff", + "choice_c": "sendemail", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8454 + }, + { + "path": "audio-1498571778.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "query", + "choice_b": "alarm_set", + "choice_c": "datetime_convert", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8455 + }, + { + "path": "audio-1502896232-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "sendemail", + "choice_b": "play_radio", + "choice_c": "calendar_remove", + "choice_d": "transport_ticket", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8456 + }, + { + "path": "audio--1505407262.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "email_sendemail", + "choice_b": "audio_volume_mute", + "choice_c": "email_querycontact", + "choice_d": "radio", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8457 + }, + { + "path": "audio-1494418135-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "social_query", + "choice_b": "datetime_query", + "choice_c": "lists_createoradd", + "choice_d": "news_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8458 + }, + { + "path": "audio-1492785178.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "lists_remove", + "choice_b": "iot_hue_lighton", + "choice_c": "qa_currency", + "choice_d": "qa_definition", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8459 + }, + { + "path": "audio-1490358908.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "music_dislikeness", + "choice_c": "iot_coffee", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8460 + }, + { + "path": "audio--1504193042-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "general_greet", + "choice_b": "play_audiobook", + "choice_c": "cooking_recipe", + "choice_d": "audio_volume_mute", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8461 + }, + { + "path": "audio-1502195021-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "alarm_set", + "choice_b": "recommendation_events", + "choice_c": "remove", + "choice_d": "lists_createoradd", + "answer_gt": "remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8462 + }, + { + "path": "audio-1488992754.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "recommendation_locations", + "choice_b": "music_likeness", + "choice_c": "play_podcasts", + "choice_d": "qa_factoid", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8463 + }, + { + "path": "audio-1498566945-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "play_music", + "choice_b": "iot_hue_lightup", + "choice_c": "calendar_remove", + "choice_d": "play_audiobook", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8464 + }, + { + "path": "audio-1488989319.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "lists_remove", + "choice_b": "general_joke", + "choice_c": "calendar_set", + "choice_d": "music_likeness", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8465 + }, + { + "path": "audio-1501685751-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "play_radio", + "choice_b": "takeaway_order", + "choice_c": "iot_hue_lightup", + "choice_d": "music_likeness", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8466 + }, + { + "path": "audio-1497271580.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "social_post", + "choice_b": "lists_remove", + "choice_c": "remove", + "choice_d": "alarm_remove", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8467 + }, + { + "path": "audio-1499692407.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "radio", + "choice_b": "audio_volume_up", + "choice_c": "play_music", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8468 + }, + { + "path": "audio-1501751396-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "transport_traffic", + "choice_b": "weather_query", + "choice_c": "hue_lightoff", + "choice_d": "radio", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8469 + }, + { + "path": "audio-1501756768.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "audio_volume_down", + "choice_b": "iot_hue_lightup", + "choice_c": "qa_definition", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8470 + }, + { + "path": "audio-1498485416-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "play_audiobook", + "choice_b": "play_music", + "choice_c": "lists_remove", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8471 + }, + { + "path": "audio-1501767220.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "createoradd", + "choice_b": "recommendation_movies", + "choice_c": "qa_factoid", + "choice_d": "transport_ticket", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8472 + }, + { + "path": "audio-1502462027.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "radio", + "choice_b": "email_query", + "choice_c": "cooking_recipe", + "choice_d": "general_joke", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8473 + }, + { + "path": "audio-1490705722.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "qa_definition", + "choice_b": "weather_query", + "choice_c": "datetime_query", + "choice_d": "play_podcasts", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8474 + }, + { + "path": "audio-1490007090-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "recommendation_events", + "choice_b": "transport_taxi", + "choice_c": "iot_coffee", + "choice_d": "remove", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8475 + }, + { + "path": "audio-1490192690.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "lists_createoradd", + "choice_c": "weather_query", + "choice_d": "query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8476 + }, + { + "path": "audio-1502099454-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "play_game", + "choice_b": "weather_query", + "choice_c": "sendemail", + "choice_d": "recommendation_locations", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8477 + }, + { + "path": "audio-1488557290.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "play_podcasts", + "choice_b": "play_radio", + "choice_c": "qa_currency", + "choice_d": "transport_taxi", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8478 + }, + { + "path": "audio-1501776505-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "weather_query", + "choice_b": "iot_cleaning", + "choice_c": "social_query", + "choice_d": "alarm_remove", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8479 + }, + { + "path": "audio-1496265704.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "iot_coffee", + "choice_b": "lists_remove", + "choice_c": "alarm_query", + "choice_d": "music_query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8480 + }, + { + "path": "audio-1502309214-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "transport_taxi", + "choice_b": "iot_hue_lightchange", + "choice_c": "datetime_query", + "choice_d": "email_addcontact", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8481 + }, + { + "path": "audio-1502201256-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "music", + "choice_b": "play_music", + "choice_c": "audio_volume_down", + "choice_d": "email_sendemail", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8482 + }, + { + "path": "audio-1502199609-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "alarm_set", + "choice_b": "iot_coffee", + "choice_c": "calendar_set", + "choice_d": "email_sendemail", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8483 + }, + { + "path": "audio--1504193105-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "audio_volume_mute", + "choice_b": "transport_taxi", + "choice_c": "hue_lightoff", + "choice_d": "iot_cleaning", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8484 + }, + { + "path": "audio-1501687561.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "alarm_set", + "choice_b": "calendar_query", + "choice_c": "email_sendemail", + "choice_d": "query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8485 + }, + { + "path": "audio-1501690027.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "news_query", + "choice_b": "radio", + "choice_c": "recommendation_events", + "choice_d": "hue_lightoff", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8486 + }, + { + "path": "audio-1501777228.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "qa_stock", + "choice_b": "weather_query", + "choice_c": "qa_factoid", + "choice_d": "music_dislikeness", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8487 + }, + { + "path": "audio-1434542021-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "sendemail", + "choice_b": "music_settings", + "choice_c": "qa_currency", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8488 + }, + { + "path": "audio-1488989231.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "lists_remove", + "choice_b": "general_joke", + "choice_c": "datetime_convert", + "choice_d": "iot_hue_lightchange", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8489 + }, + { + "path": "audio-1490366492-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "music_settings", + "choice_b": "cooking_recipe", + "choice_c": "recommendation_locations", + "choice_d": "remove", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8490 + }, + { + "path": "audio-1490291461.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "general_quirky", + "choice_b": "calendar_set", + "choice_c": "weather_query", + "choice_d": "qa_definition", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8491 + }, + { + "path": "audio-1490290958.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "joke", + "choice_b": "transport_traffic", + "choice_c": "sendemail", + "choice_d": "transport_query", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8492 + }, + { + "path": "audio-1497884188-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "post", + "choice_b": "music_query", + "choice_c": "createoradd", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8493 + }, + { + "path": "audio-1490109056.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_wemo_on", + "choice_b": "play_game", + "choice_c": "play_radio", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8494 + }, + { + "path": "audio-1490263672-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "social_post", + "choice_b": "play_music", + "choice_c": "email_addcontact", + "choice_d": "iot_cleaning", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8495 + }, + { + "path": "audio--1504197622.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "music_dislikeness", + "choice_b": "iot_wemo_on", + "choice_c": "play_audiobook", + "choice_d": "alarm_remove", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8496 + }, + { + "path": "audio-1499697296.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "calendar_set", + "choice_b": "play_radio", + "choice_c": "play_game", + "choice_d": "music_likeness", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8497 + }, + { + "path": "audio-1497862081.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "transport_query", + "choice_b": "alarm_set", + "choice_c": "weather_query", + "choice_d": "query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8498 + }, + { + "path": "audio-1497862912-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_hue_lightchange", + "choice_b": "query", + "choice_c": "transport_taxi", + "choice_d": "transport_traffic", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8499 + }, + { + "path": "audio-1490369046-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "sendemail", + "choice_b": "audio_volume_down", + "choice_c": "takeaway_order", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8500 + }, + { + "path": "audio-1490017428-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "lists_query", + "choice_b": "transport_query", + "choice_c": "music_settings", + "choice_d": "alarm_remove", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8501 + }, + { + "path": "audio-1498483119-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "qa_definition", + "choice_c": "cooking_recipe", + "choice_d": "calendar_remove", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8502 + }, + { + "path": "audio-1489153408.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "lists_query", + "choice_b": "recommendation_events", + "choice_c": "iot_cleaning", + "choice_d": "lists_remove", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8503 + }, + { + "path": "audio-1497270676.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "music_settings", + "choice_b": "lists_createoradd", + "choice_c": "cooking_recipe", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8504 + }, + { + "path": "audio-1501691937-headset.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "audio_volume_down", + "choice_b": "calendar_remove", + "choice_c": "news_query", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8505 + }, + { + "path": "audio-1501752800-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "music", + "choice_b": "qa_currency", + "choice_c": "calendar_remove", + "choice_d": "locations", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8506 + }, + { + "path": "audio-1497885924-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "general_quirky", + "choice_b": "lists_createoradd", + "choice_c": "iot_hue_lightup", + "choice_d": "recommendation_locations", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8507 + }, + { + "path": "audio-1497024759-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "general_joke", + "choice_b": "transport_query", + "choice_c": "play_game", + "choice_d": "iot_hue_lightchange", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8508 + }, + { + "path": "audio-1502891080-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "alarm_remove", + "choice_b": "email_addcontact", + "choice_c": "iot_coffee", + "choice_d": "lists_remove", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8509 + }, + { + "path": "audio-1502197579.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "general_greet", + "choice_b": "iot_hue_lightoff", + "choice_c": "calendar_set", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8510 + }, + { + "path": "audio-1490022101.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "remove", + "choice_b": "calendar_remove", + "choice_c": "email_addcontact", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8511 + }, + { + "path": "audio-1502896214.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "calendar_query", + "choice_b": "social_query", + "choice_c": "play_game", + "choice_d": "play_music", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8512 + }, + { + "path": "audio-1501157521-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "iot_wemo_on", + "choice_b": "email_addcontact", + "choice_c": "cooking_recipe", + "choice_d": "lists_createoradd", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8513 + }, + { + "path": "audio-1499265642.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "joke", + "choice_c": "general_greet", + "choice_d": "qa_definition", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8514 + }, + { + "path": "audio-1434529654-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "music_dislikeness", + "choice_c": "lists_createoradd", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8515 + }, + { + "path": "audio-1494548237.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "general_joke", + "choice_b": "play_game", + "choice_c": "datetime_query", + "choice_d": "calendar_remove", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8516 + }, + { + "path": "audio-1502100312.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "transport_traffic", + "choice_b": "iot_hue_lightdim", + "choice_c": "calendar_remove", + "choice_d": "email_querycontact", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8517 + }, + { + "path": "audio-1434540137.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "email_sendemail", + "choice_b": "iot_hue_lightup", + "choice_c": "general_quirky", + "choice_d": "lists_createoradd", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8518 + }, + { + "path": "audio-1502308541-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "play_audiobook", + "choice_b": "email_query", + "choice_c": "calendar_set", + "choice_d": "audio_volume_down", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8519 + }, + { + "path": "audio-1434535075-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "recommendation_movies", + "choice_b": "qa_factoid", + "choice_c": "email_sendemail", + "choice_d": "social_query", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8520 + }, + { + "path": "audio-1499267031.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "alarm_query", + "choice_b": "createoradd", + "choice_c": "music_query", + "choice_d": "audio_volume_up", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8521 + }, + { + "path": "audio-1497270940.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "play_music", + "choice_b": "iot_wemo_on", + "choice_c": "iot_hue_lightchange", + "choice_d": "weather_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8522 + }, + { + "path": "audio-1502376209.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "recommendation_events", + "choice_b": "calendar_set", + "choice_c": "takeaway_query", + "choice_d": "play_audiobook", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8523 + }, + { + "path": "audio-1501692963.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "general_greet", + "choice_b": "play_game", + "choice_c": "iot_hue_lightchange", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8524 + }, + { + "path": "audio-1502897050-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "audio_volume_mute", + "choice_b": "play_music", + "choice_c": "audio_volume_up", + "choice_d": "email_sendemail", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8525 + }, + { + "path": "audio-1490283095.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "social_query", + "choice_b": "lists_query", + "choice_c": "iot_hue_lightup", + "choice_d": "music", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8526 + }, + { + "path": "audio-1497883378.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "iot_hue_lightchange", + "choice_b": "iot_hue_lightdim", + "choice_c": "music_likeness", + "choice_d": "cooking_recipe", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8527 + }, + { + "path": "audio-1501778448-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "hue_lightdim", + "choice_b": "general_joke", + "choice_c": "play_radio", + "choice_d": "email_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8528 + }, + { + "path": "audio-1498566663.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "calendar_remove", + "choice_b": "query", + "choice_c": "transport_taxi", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8529 + }, + { + "path": "audio-1501406713.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_taxi", + "choice_b": "music_likeness", + "choice_c": "social_post", + "choice_d": "recommendation_movies", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8530 + }, + { + "path": "audio-1490359047-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "radio", + "choice_b": "calendar_remove", + "choice_c": "iot_hue_lightdim", + "choice_d": "cooking_recipe", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8531 + }, + { + "path": "audio-1497270000-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "general_greet", + "choice_b": "qa_definition", + "choice_c": "music", + "choice_d": "iot_hue_lightup", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8532 + }, + { + "path": "audio-1499689999-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "calendar_set", + "choice_b": "email_addcontact", + "choice_c": "play_audiobook", + "choice_d": "qa_factoid", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8533 + }, + { + "path": "audio-1501775422-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "qa_currency", + "choice_b": "post", + "choice_c": "email_sendemail", + "choice_d": "calendar_remove", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8534 + }, + { + "path": "audio--1504193014.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "alarm_query", + "choice_b": "play_audiobook", + "choice_c": "social_post", + "choice_d": "iot_hue_lightoff", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8535 + }, + { + "path": "audio-1499689440.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_factoid", + "choice_b": "iot_wemo_on", + "choice_c": "iot_hue_lightchange", + "choice_d": "radio", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8536 + }, + { + "path": "audio-1502890178-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "remove", + "choice_b": "audio_volume_up", + "choice_c": "alarm_set", + "choice_d": "recommendation_movies", + "answer_gt": "audio_volume_up", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8537 + }, + { + "path": "audio-1502309711.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "iot_hue_lightoff", + "choice_b": "music", + "choice_c": "calendar_remove", + "choice_d": "iot_cleaning", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8538 + }, + { + "path": "audio-1488622044.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "iot_coffee", + "choice_b": "audio_volume_other", + "choice_c": "play_radio", + "choice_d": "transport_traffic", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8539 + }, + { + "path": "audio-1488563776.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "calendar_set", + "choice_b": "email_sendemail", + "choice_c": "takeaway_order", + "choice_d": "music_likeness", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8540 + }, + { + "path": "audio-1497886758.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "music_settings", + "choice_b": "news_query", + "choice_c": "calendar_remove", + "choice_d": "transport_taxi", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8541 + }, + { + "path": "audio-1501765622-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "alarm_set", + "choice_b": "qa_currency", + "choice_c": "calendar_remove", + "choice_d": "music_query", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8542 + }, + { + "path": "audio-1497872816-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "weather_query", + "choice_b": "post", + "choice_c": "iot_hue_lighton", + "choice_d": "alarm_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8543 + }, + { + "path": "audio-1489153316.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "general_joke", + "choice_b": "general_quirky", + "choice_c": "play_audiobook", + "choice_d": "iot_hue_lightdim", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8544 + }, + { + "path": "audio-1501764451.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "qa_factoid", + "choice_b": "calendar_remove", + "choice_c": "query", + "choice_d": "play_audiobook", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8545 + }, + { + "path": "audio-1501158522.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "general_joke", + "choice_b": "music_dislikeness", + "choice_c": "createoradd", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8546 + }, + { + "path": "audio-1499267431-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "sendemail", + "choice_b": "transport_taxi", + "choice_c": "recommendation_movies", + "choice_d": "recommendation_locations", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8547 + }, + { + "path": "audio-1490954828.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "music_likeness", + "choice_b": "social_query", + "choice_c": "transport_taxi", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8548 + }, + { + "path": "audio-1502100016.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "qa_maths", + "choice_b": "recommendation_locations", + "choice_c": "iot_wemo_on", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8549 + }, + { + "path": "audio-1498758847-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "iot_hue_lightchange", + "choice_b": "qa_factoid", + "choice_c": "takeaway_query", + "choice_d": "play_audiobook", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8550 + }, + { + "path": "audio-1492779883.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "qa_stock", + "choice_c": "qa_definition", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8551 + }, + { + "path": "audio-1502218725-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "qa_definition", + "choice_b": "transport_traffic", + "choice_c": "recommendation_locations", + "choice_d": "qa_maths", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8552 + }, + { + "path": "audio-1502199657-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "audio_volume_down", + "choice_b": "iot_hue_lightoff", + "choice_c": "iot_hue_lightup", + "choice_d": "play_radio", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8553 + }, + { + "path": "audio-1489154243.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "general_greet", + "choice_b": "recommendation_locations", + "choice_c": "calendar_remove", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8554 + }, + { + "path": "audio-1501773014-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "news_query", + "choice_b": "general_quirky", + "choice_c": "locations", + "choice_d": "music_settings", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8555 + }, + { + "path": "audio-1502216069-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "recommendation_events", + "choice_b": "takeaway_query", + "choice_c": "play_music", + "choice_d": "locations", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8556 + }, + { + "path": "audio-1497622341.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "calendar_query", + "choice_c": "hue_lightoff", + "choice_d": "general_quirky", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8557 + }, + { + "path": "audio-1490291856-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "play_music", + "choice_b": "music_query", + "choice_c": "hue_lightdim", + "choice_d": "radio", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8558 + }, + { + "path": "audio-1501752628.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "hue_lightdim", + "choice_b": "email_query", + "choice_c": "iot_wemo_off", + "choice_d": "audio_volume_up", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8559 + }, + { + "path": "audio--1504191344-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "qa_currency", + "choice_b": "play_audiobook", + "choice_c": "transport_taxi", + "choice_d": "alarm_remove", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8560 + }, + { + "path": "audio-1494608777.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "qa_factoid", + "choice_b": "iot_wemo_on", + "choice_c": "alarm_query", + "choice_d": "play_game", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8561 + }, + { + "path": "audio-1501768107-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "locations", + "choice_b": "music_likeness", + "choice_c": "recommendation_locations", + "choice_d": "joke", + "answer_gt": "joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8562 + }, + { + "path": "audio-1490019568-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "play_music", + "choice_b": "transport_traffic", + "choice_c": "calendar_remove", + "choice_d": "email_sendemail", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8563 + }, + { + "path": "audio-1490288716-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "qa_definition", + "choice_b": "audio_volume_other", + "choice_c": "audio_volume_mute", + "choice_d": "weather_query", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8564 + }, + { + "path": "audio-1490370194.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "radio", + "choice_b": "recommendation_locations", + "choice_c": "general_quirky", + "choice_d": "locations", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8565 + }, + { + "path": "audio-1495372243-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "hue_lightdim", + "choice_b": "qa_definition", + "choice_c": "play_audiobook", + "choice_d": "iot_hue_lighton", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8566 + }, + { + "path": "audio-1502894611-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "social_query", + "choice_b": "qa_currency", + "choice_c": "iot_hue_lightdim", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8567 + }, + { + "path": "audio-1501777688.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "play_music", + "choice_b": "lists_remove", + "choice_c": "transport_ticket", + "choice_d": "audio_volume_down", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8568 + }, + { + "path": "audio-1499695912-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "datetime_convert", + "choice_b": "iot_cleaning", + "choice_c": "general_quirky", + "choice_d": "query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8569 + }, + { + "path": "audio-1489498486.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "music_settings", + "choice_c": "lists_query", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8570 + }, + { + "path": "audio-1501752238.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "email_query", + "choice_b": "music_likeness", + "choice_c": "email_addcontact", + "choice_d": "sendemail", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8571 + }, + { + "path": "audio-1501406964-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "general_quirky", + "choice_b": "music_likeness", + "choice_c": "alarm_set", + "choice_d": "calendar_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8572 + }, + { + "path": "audio-1494512809.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "calendar_remove", + "choice_b": "createoradd", + "choice_c": "audio_volume_mute", + "choice_d": "iot_hue_lightup", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8573 + }, + { + "path": "audio-1499693817.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "play_audiobook", + "choice_b": "music_likeness", + "choice_c": "email_querycontact", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8574 + }, + { + "path": "audio-1497882593.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "calendar_set", + "choice_b": "hue_lightoff", + "choice_c": "iot_hue_lightdim", + "choice_d": "cooking_recipe", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8575 + }, + { + "path": "audio--1505405413.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "play_audiobook", + "choice_b": "calendar_query", + "choice_c": "alarm_set", + "choice_d": "datetime_convert", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8576 + }, + { + "path": "audio--1504190327.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "social_query", + "choice_b": "music_query", + "choice_c": "datetime_query", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8577 + }, + { + "path": "audio-1502200784-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "calendar_set", + "choice_b": "transport_ticket", + "choice_c": "email_query", + "choice_d": "recommendation_movies", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8578 + }, + { + "path": "audio-1502894308.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "music", + "choice_b": "music_query", + "choice_c": "qa_definition", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8579 + }, + { + "path": "audio-1495733739-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "news_query", + "choice_b": "iot_hue_lightup", + "choice_c": "cooking_recipe", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8580 + }, + { + "path": "audio-1501758154-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "music_likeness", + "choice_b": "remove", + "choice_c": "alarm_query", + "choice_d": "locations", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8581 + }, + { + "path": "audio--1504191551.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "alarm_remove", + "choice_b": "music_likeness", + "choice_c": "play_podcasts", + "choice_d": "qa_factoid", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8582 + }, + { + "path": "audio-1490286503-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "play_radio", + "choice_b": "audio_volume_down", + "choice_c": "audio_volume_other", + "choice_d": "iot_hue_lightdim", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8583 + }, + { + "path": "audio-1502199179-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "recommendation_locations", + "choice_b": "datetime_query", + "choice_c": "qa_maths", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8584 + }, + { + "path": "audio-1490357449.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "audio_volume_mute", + "choice_b": "audio_volume_up", + "choice_c": "calendar_set", + "choice_d": "news_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8585 + }, + { + "path": "audio-1502195103-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "locations", + "choice_b": "joke", + "choice_c": "createoradd", + "choice_d": "hue_lightdim", + "answer_gt": "hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8586 + }, + { + "path": "audio-1490955830-headset.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "calendar_set", + "choice_b": "transport_ticket", + "choice_c": "social_post", + "choice_d": "remove", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8587 + }, + { + "path": "audio-1502218365-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "qa_maths", + "choice_b": "recommendation_locations", + "choice_c": "transport_query", + "choice_d": "iot_hue_lightdim", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8588 + }, + { + "path": "audio-1497555247-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "qa_maths", + "choice_b": "cooking_recipe", + "choice_c": "joke", + "choice_d": "datetime_query", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8589 + }, + { + "path": "audio-1501172981.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "iot_cleaning", + "choice_b": "query", + "choice_c": "takeaway_order", + "choice_d": "recommendation_events", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8590 + }, + { + "path": "audio--1505406003.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "recommendation_locations", + "choice_b": "email_query", + "choice_c": "music", + "choice_d": "locations", + "answer_gt": "music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8591 + }, + { + "path": "audio-1497886770.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "query", + "choice_b": "joke", + "choice_c": "news_query", + "choice_d": "general_greet", + "answer_gt": "joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8592 + }, + { + "path": "audio-1490366428-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "iot_hue_lighton", + "choice_b": "iot_hue_lightup", + "choice_c": "qa_factoid", + "choice_d": "play_audiobook", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8593 + }, + { + "path": "audio-1501691010.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "qa_currency", + "choice_b": "audio_volume_mute", + "choice_c": "email_querycontact", + "choice_d": "lists_createoradd", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8594 + }, + { + "path": "audio-1488562994.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "cooking_recipe", + "choice_b": "audio_volume_mute", + "choice_c": "music_likeness", + "choice_d": "play_radio", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8595 + }, + { + "path": "audio-1497557395.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "calendar_query", + "choice_b": "createoradd", + "choice_c": "iot_cleaning", + "choice_d": "weather_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8596 + }, + { + "path": "audio--1504190472-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_cleaning", + "choice_b": "email_querycontact", + "choice_c": "news_query", + "choice_d": "audio_volume_down", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8597 + }, + { + "path": "audio-1490359679.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "query", + "choice_b": "play_radio", + "choice_c": "transport_ticket", + "choice_d": "transport_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8598 + }, + { + "path": "audio--1505405672-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "qa_definition", + "choice_b": "general_joke", + "choice_c": "qa_factoid", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8599 + }, + { + "path": "audio--1506079789-headset.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "alarm_remove", + "choice_b": "createoradd", + "choice_c": "hue_lightoff", + "choice_d": "iot_hue_lightdim", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8600 + }, + { + "path": "audio-1500038684-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "email_query", + "choice_b": "recommendation_movies", + "choice_c": "query", + "choice_d": "takeaway_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8601 + }, + { + "path": "audio-1497619657-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "qa_maths", + "choice_b": "calendar_set", + "choice_c": "iot_cleaning", + "choice_d": "qa_currency", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8602 + }, + { + "path": "audio-1501755847.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "lists_createoradd", + "choice_b": "lists_remove", + "choice_c": "iot_hue_lightchange", + "choice_d": "play_audiobook", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8603 + }, + { + "path": "audio-1490008047.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "music", + "choice_b": "createoradd", + "choice_c": "email_sendemail", + "choice_d": "iot_wemo_on", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8604 + }, + { + "path": "audio--1505406920-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "remove", + "choice_b": "transport_ticket", + "choice_c": "recommendation_movies", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8605 + }, + { + "path": "audio-1501754231-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "music_settings", + "choice_b": "cooking_recipe", + "choice_c": "recommendation_events", + "choice_d": "iot_hue_lightchange", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8606 + }, + { + "path": "audio-1490022261-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "music", + "choice_b": "transport_ticket", + "choice_c": "play_game", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8607 + }, + { + "path": "audio-1497865219-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "takeaway_order", + "choice_b": "joke", + "choice_c": "lists_createoradd", + "choice_d": "general_greet", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8608 + }, + { + "path": "audio-1498483490-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "calendar_set", + "choice_b": "qa_maths", + "choice_c": "play_music", + "choice_d": "alarm_set", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8609 + }, + { + "path": "audio-1434533768-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "email_query", + "choice_b": "play_audiobook", + "choice_c": "iot_hue_lightdim", + "choice_d": "recommendation_events", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8610 + }, + { + "path": "audio-1488993620.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "joke", + "choice_b": "play_podcasts", + "choice_c": "calendar_query", + "choice_d": "qa_maths", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8611 + }, + { + "path": "audio-1502377700-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "play_radio", + "choice_b": "general_joke", + "choice_c": "alarm_set", + "choice_d": "iot_hue_lightoff", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8612 + }, + { + "path": "audio-1490261505-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "qa_currency", + "choice_b": "createoradd", + "choice_c": "radio", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8613 + }, + { + "path": "audio-1498571075-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "audio_volume_down", + "choice_b": "play_radio", + "choice_c": "social_post", + "choice_d": "transport_query", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8614 + }, + { + "path": "audio-1501774631.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "iot_cleaning", + "choice_c": "alarm_query", + "choice_d": "remove", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8615 + }, + { + "path": "audio-1494608777-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "iot_hue_lightchange", + "choice_b": "social_post", + "choice_c": "takeaway_order", + "choice_d": "play_game", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8616 + }, + { + "path": "audio-1497450417.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "play_radio", + "choice_b": "qa_definition", + "choice_c": "play_music", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8617 + }, + { + "path": "audio--1504193348-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "cooking_recipe", + "choice_b": "iot_wemo_on", + "choice_c": "audio_volume_mute", + "choice_d": "lists_remove", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8618 + }, + { + "path": "audio-1501690018.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "datetime_query", + "choice_b": "joke", + "choice_c": "transport_taxi", + "choice_d": "social_query", + "answer_gt": "joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8619 + }, + { + "path": "audio-1502113042.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "play_game", + "choice_b": "qa_factoid", + "choice_c": "datetime_query", + "choice_d": "lists_createoradd", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8620 + }, + { + "path": "audio-1498578513.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "hue_lightdim", + "choice_b": "calendar_query", + "choice_c": "play_game", + "choice_d": "lists_createoradd", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8621 + }, + { + "path": "audio-1501686151-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "audio_volume_up", + "choice_b": "general_greet", + "choice_c": "lists_query", + "choice_d": "qa_maths", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8622 + }, + { + "path": "audio--1504194459.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "play_audiobook", + "choice_b": "audio_volume_down", + "choice_c": "iot_cleaning", + "choice_d": "transport_ticket", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8623 + }, + { + "path": "audio-1490194380.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "alarm_set", + "choice_b": "alarm_remove", + "choice_c": "qa_definition", + "choice_d": "music_settings", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8624 + }, + { + "path": "audio-1501697393.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "play_radio", + "choice_b": "social_post", + "choice_c": "calendar_set", + "choice_d": "iot_hue_lightup", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8625 + }, + { + "path": "audio-1498577116.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "email_query", + "choice_b": "iot_coffee", + "choice_c": "music_settings", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8626 + }, + { + "path": "audio-1488557223.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "social_post", + "choice_b": "email_query", + "choice_c": "recommendation_locations", + "choice_d": "query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8627 + }, + { + "path": "audio-1434531688-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "iot_hue_lightoff", + "choice_b": "general_greet", + "choice_c": "iot_cleaning", + "choice_d": "hue_lightdim", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8628 + }, + { + "path": "audio-1498485256.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "iot_hue_lightoff", + "choice_b": "email_query", + "choice_c": "iot_hue_lightup", + "choice_d": "cooking_recipe", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8629 + }, + { + "path": "audio-1494607257-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "iot_coffee", + "choice_b": "audio_volume_up", + "choice_c": "music_likeness", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8630 + }, + { + "path": "audio-1500388165-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "datetime_query", + "choice_b": "recommendation_events", + "choice_c": "iot_hue_lightoff", + "choice_d": "iot_wemo_off", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8631 + }, + { + "path": "audio-1489498421.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "calendar_query", + "choice_b": "play_game", + "choice_c": "qa_definition", + "choice_d": "calendar_remove", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8632 + }, + { + "path": "audio-1434538104.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "remove", + "choice_b": "email_querycontact", + "choice_c": "takeaway_order", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8633 + }, + { + "path": "audio--1504190658.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "createoradd", + "choice_b": "general_greet", + "choice_c": "hue_lightdim", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8634 + }, + { + "path": "audio-1501406519-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "transport_traffic", + "choice_b": "datetime_query", + "choice_c": "social_query", + "choice_d": "music", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8635 + }, + { + "path": "audio-1495369698-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "lists_remove", + "choice_b": "iot_wemo_off", + "choice_c": "audio_volume_mute", + "choice_d": "hue_lightoff", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8636 + }, + { + "path": "audio-1490195363-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "play_podcasts", + "choice_b": "play_music", + "choice_c": "transport_traffic", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8637 + }, + { + "path": "audio-1501687727.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "audio_volume_other", + "choice_c": "play_game", + "choice_d": "qa_definition", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8638 + }, + { + "path": "audio-1490008945.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "recommendation_events", + "choice_b": "social_post", + "choice_c": "music_likeness", + "choice_d": "music_dislikeness", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8639 + }, + { + "path": "audio-1499690724-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "audio_volume_down", + "choice_b": "music_settings", + "choice_c": "iot_wemo_off", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8640 + }, + { + "path": "audio-1497887650.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "calendar_query", + "choice_b": "music_dislikeness", + "choice_c": "alarm_remove", + "choice_d": "iot_hue_lightchange", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8641 + }, + { + "path": "audio-1490955368-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "recommendation_locations", + "choice_b": "email_sendemail", + "choice_c": "general_quirky", + "choice_d": "datetime_convert", + "answer_gt": "recommendation_locations", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8642 + }, + { + "path": "audio-1497884418.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "lists_remove", + "choice_b": "lists_query", + "choice_c": "news_query", + "choice_d": "calendar_remove", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8643 + }, + { + "path": "audio-1490182883-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "music_likeness", + "choice_b": "email_query", + "choice_c": "alarm_remove", + "choice_d": "iot_coffee", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8644 + }, + { + "path": "audio-1501777061.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "news_query", + "choice_b": "music_likeness", + "choice_c": "qa_factoid", + "choice_d": "play_game", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8645 + }, + { + "path": "audio-1490021134.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "iot_hue_lightdim", + "choice_b": "takeaway_order", + "choice_c": "calendar_remove", + "choice_d": "cooking_recipe", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8646 + }, + { + "path": "audio-1494422710-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "post", + "choice_b": "play_game", + "choice_c": "locations", + "choice_d": "iot_hue_lighton", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8647 + }, + { + "path": "audio-1489153917.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "qa_factoid", + "choice_b": "iot_hue_lightup", + "choice_c": "general_greet", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8648 + }, + { + "path": "audio-1490183133.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "query", + "choice_b": "transport_ticket", + "choice_c": "iot_cleaning", + "choice_d": "iot_hue_lightup", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8649 + }, + { + "path": "audio-1500898226.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "qa_stock", + "choice_b": "play_audiobook", + "choice_c": "music_query", + "choice_d": "recommendation_locations", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8650 + }, + { + "path": "audio-1498573960-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "iot_cleaning", + "choice_b": "music_settings", + "choice_c": "lists_createoradd", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8651 + }, + { + "path": "audio-1502895701.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "iot_hue_lighton", + "choice_b": "iot_hue_lightchange", + "choice_c": "alarm_set", + "choice_d": "audio_volume_down", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8652 + }, + { + "path": "audio-1494417438.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "radio", + "choice_b": "play_radio", + "choice_c": "recommendation_events", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8653 + }, + { + "path": "audio-1498485352.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "email_sendemail", + "choice_b": "sendemail", + "choice_c": "iot_coffee", + "choice_d": "audio_volume_other", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8654 + }, + { + "path": "audio-1502200900-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "calendar_query", + "choice_b": "transport_query", + "choice_c": "general_joke", + "choice_d": "music_likeness", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8655 + }, + { + "path": "audio-1502896185-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "transport_traffic", + "choice_b": "transport_taxi", + "choice_c": "calendar_set", + "choice_d": "transport_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8656 + }, + { + "path": "audio-1497886910-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "iot_hue_lightoff", + "choice_b": "calendar_set", + "choice_c": "music_query", + "choice_d": "locations", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8657 + }, + { + "path": "audio-1502299618.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "transport_ticket", + "choice_b": "email_querycontact", + "choice_c": "recommendation_movies", + "choice_d": "transport_traffic", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8658 + }, + { + "path": "audio-1497881782.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "audio_volume_other", + "choice_b": "cooking_recipe", + "choice_c": "qa_maths", + "choice_d": "qa_stock", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8659 + }, + { + "path": "audio--1505404136-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "iot_wemo_on", + "choice_b": "datetime_query", + "choice_c": "sendemail", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8660 + }, + { + "path": "audio--1504191382.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "music_likeness", + "choice_b": "audio_volume_mute", + "choice_c": "recommendation_locations", + "choice_d": "iot_cleaning", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8661 + }, + { + "path": "audio-1501753190-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "audio_volume_other", + "choice_b": "lists_query", + "choice_c": "transport_ticket", + "choice_d": "qa_definition", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8662 + }, + { + "path": "audio-1501752883.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "remove", + "choice_b": "iot_hue_lightoff", + "choice_c": "play_radio", + "choice_d": "sendemail", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8663 + }, + { + "path": "audio-1502197869-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "music_query", + "choice_b": "iot_hue_lightdim", + "choice_c": "hue_lightdim", + "choice_d": "query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8664 + }, + { + "path": "audio--1504192388.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "datetime_convert", + "choice_b": "query", + "choice_c": "alarm_set", + "choice_d": "general_greet", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8665 + }, + { + "path": "audio-1501757123.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "music_dislikeness", + "choice_b": "recommendation_locations", + "choice_c": "qa_factoid", + "choice_d": "radio", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8666 + }, + { + "path": "audio--1504192812.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "general_quirky", + "choice_b": "audio_volume_mute", + "choice_c": "audio_volume_up", + "choice_d": "takeaway_order", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8667 + }, + { + "path": "audio-1490957105-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "audio_volume_other", + "choice_b": "news_query", + "choice_c": "calendar_set", + "choice_d": "takeaway_query", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8668 + }, + { + "path": "audio-1501777083.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "cooking_recipe", + "choice_b": "lists_remove", + "choice_c": "iot_hue_lighton", + "choice_d": "music_query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8669 + }, + { + "path": "audio-1498646783.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "qa_definition", + "choice_b": "query", + "choice_c": "iot_coffee", + "choice_d": "transport_traffic", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8670 + }, + { + "path": "audio-1498566346.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "recommendation_events", + "choice_b": "social_query", + "choice_c": "music_likeness", + "choice_d": "alarm_remove", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8671 + }, + { + "path": "audio-1489154716.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "email_addcontact", + "choice_b": "iot_wemo_off", + "choice_c": "general_joke", + "choice_d": "play_music", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8672 + }, + { + "path": "audio-1502101978.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "takeaway_order", + "choice_b": "social_query", + "choice_c": "iot_wemo_on", + "choice_d": "qa_currency", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8673 + }, + { + "path": "audio-1495371776-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "calendar_query", + "choice_b": "general_greet", + "choice_c": "music_dislikeness", + "choice_d": "sendemail", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8674 + }, + { + "path": "audio-1501754980.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "iot_coffee", + "choice_b": "qa_factoid", + "choice_c": "lists_remove", + "choice_d": "play_music", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8675 + }, + { + "path": "audio-1501150748.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "lists_query", + "choice_b": "sendemail", + "choice_c": "iot_cleaning", + "choice_d": "audio_volume_up", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8676 + }, + { + "path": "audio-1490106100.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "general_quirky", + "choice_b": "calendar_query", + "choice_c": "music_likeness", + "choice_d": "qa_currency", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8677 + }, + { + "path": "audio-1490359951-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "alarm_set", + "choice_b": "audio_volume_down", + "choice_c": "qa_stock", + "choice_d": "lists_createoradd", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8678 + }, + { + "path": "audio-1502302100.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "iot_wemo_off", + "choice_b": "alarm_remove", + "choice_c": "alarm_query", + "choice_d": "email_query", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8679 + }, + { + "path": "audio-1434538853-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "email_querycontact", + "choice_b": "general_greet", + "choice_c": "recommendation_locations", + "choice_d": "calendar_query", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8680 + }, + { + "path": "audio-1494421392.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "datetime_query", + "choice_b": "general_quirky", + "choice_c": "post", + "choice_d": "query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8681 + }, + { + "path": "audio-1498565143.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "recommendation_events", + "choice_b": "general_quirky", + "choice_c": "locations", + "choice_d": "alarm_set", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8682 + }, + { + "path": "audio-1490955539-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "email_addcontact", + "choice_b": "play_radio", + "choice_c": "news_query", + "choice_d": "audio_volume_down", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8683 + }, + { + "path": "audio-1497555701.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "play_podcasts", + "choice_b": "email_addcontact", + "choice_c": "news_query", + "choice_d": "takeaway_order", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8684 + }, + { + "path": "audio-1498485165-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "social_query", + "choice_b": "calendar_query", + "choice_c": "iot_hue_lightoff", + "choice_d": "iot_cleaning", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8685 + }, + { + "path": "audio-1498566094-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "createoradd", + "choice_b": "lists_createoradd", + "choice_c": "datetime_convert", + "choice_d": "qa_currency", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8686 + }, + { + "path": "audio-1502197310.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "alarm_set", + "choice_c": "weather_query", + "choice_d": "social_post", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8687 + }, + { + "path": "audio-1490199719-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "alarm_query", + "choice_b": "calendar_set", + "choice_c": "joke", + "choice_d": "social_post", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8688 + }, + { + "path": "audio--1505324888-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "lists_createoradd", + "choice_b": "remove", + "choice_c": "calendar_query", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8689 + }, + { + "path": "audio-1501150765.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "calendar_set", + "choice_b": "iot_hue_lightchange", + "choice_c": "social_post", + "choice_d": "takeaway_order", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8690 + }, + { + "path": "audio-1501771665-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "email_querycontact", + "choice_b": "general_greet", + "choice_c": "alarm_query", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8691 + }, + { + "path": "audio-1502200898.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "datetime_query", + "choice_b": "transport_query", + "choice_c": "radio", + "choice_d": "play_audiobook", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8692 + }, + { + "path": "audio--1505405343.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "qa_factoid", + "choice_b": "iot_hue_lightup", + "choice_c": "iot_coffee", + "choice_d": "audio_volume_mute", + "answer_gt": "audio_volume_mute", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8693 + }, + { + "path": "audio-1501693130.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "social_post", + "choice_b": "play_radio", + "choice_c": "lists_remove", + "choice_d": "general_joke", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8694 + }, + { + "path": "audio-1489497860.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "alarm_remove", + "choice_b": "alarm_set", + "choice_c": "play_radio", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8695 + }, + { + "path": "audio-1501158031.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "recommendation_movies", + "choice_b": "qa_maths", + "choice_c": "radio", + "choice_d": "general_quirky", + "answer_gt": "recommendation_movies", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8696 + }, + { + "path": "audio-1497029701-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "lists_createoradd", + "choice_b": "locations", + "choice_c": "query", + "choice_d": "remove", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8697 + }, + { + "path": "audio-1492779900-headset.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "alarm_query", + "choice_b": "iot_hue_lightchange", + "choice_c": "audio_volume_down", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8698 + }, + { + "path": "audio-1490022019.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "takeaway_query", + "choice_b": "audio_volume_down", + "choice_c": "weather_query", + "choice_d": "qa_factoid", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8699 + }, + { + "path": "audio-1502111935-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "email_query", + "choice_b": "radio", + "choice_c": "play_podcasts", + "choice_d": "createoradd", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8700 + }, + { + "path": "audio-1490281122.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "play_audiobook", + "choice_b": "calendar_set", + "choice_c": "music", + "choice_d": "query", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8701 + }, + { + "path": "audio-1495369440-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "datetime_query", + "choice_b": "calendar_query", + "choice_c": "iot_wemo_off", + "choice_d": "iot_hue_lightoff", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8702 + }, + { + "path": "audio-1501692440.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "lists_query", + "choice_b": "play_radio", + "choice_c": "datetime_convert", + "choice_d": "sendemail", + "answer_gt": "datetime_convert", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8703 + }, + { + "path": "audio-1494514261-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "lists_query", + "choice_b": "transport_ticket", + "choice_c": "calendar_query", + "choice_d": "locations", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8704 + }, + { + "path": "audio-1490109036.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "alarm_remove", + "choice_b": "iot_hue_lightoff", + "choice_c": "transport_traffic", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8705 + }, + { + "path": "audio--1505323813.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "general_joke", + "choice_b": "email_querycontact", + "choice_c": "news_query", + "choice_d": "email_addcontact", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8706 + }, + { + "path": "audio-1502216591-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "social_post", + "choice_b": "recommendation_events", + "choice_c": "calendar_query", + "choice_d": "lists_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8707 + }, + { + "path": "audio-1497437209.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "alarm_set", + "choice_b": "iot_hue_lightdim", + "choice_c": "datetime_query", + "choice_d": "play_podcasts", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8708 + }, + { + "path": "audio-1501758119.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "music_query", + "choice_b": "calendar_remove", + "choice_c": "play_music", + "choice_d": "lists_remove", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8709 + }, + { + "path": "audio-1490106669.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "social_post", + "choice_b": "email_query", + "choice_c": "hue_lightoff", + "choice_d": "weather_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8710 + }, + { + "path": "audio-1501778636-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "qa_currency", + "choice_b": "createoradd", + "choice_c": "hue_lightdim", + "choice_d": "query", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8711 + }, + { + "path": "audio-1501758762.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "alarm_set", + "choice_b": "datetime_query", + "choice_c": "audio_volume_other", + "choice_d": "audio_volume_up", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8712 + }, + { + "path": "audio-1489154104.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "alarm_set", + "choice_b": "email_sendemail", + "choice_c": "sendemail", + "choice_d": "play_music", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8713 + }, + { + "path": "audio-1490799897.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "play_game", + "choice_b": "music_query", + "choice_c": "cooking_recipe", + "choice_d": "lists_query", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8714 + }, + { + "path": "audio-1490261731.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "iot_cleaning", + "choice_b": "calendar_query", + "choice_c": "datetime_query", + "choice_d": "audio_volume_down", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8715 + }, + { + "path": "audio-1434533488-headset.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "cooking_recipe", + "choice_b": "play_audiobook", + "choice_c": "post", + "choice_d": "alarm_set", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8716 + }, + { + "path": "audio-1490109464-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "play_music", + "choice_b": "calendar_query", + "choice_c": "play_game", + "choice_d": "lists_remove", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8717 + }, + { + "path": "audio-1490104290.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "email_addcontact", + "choice_b": "alarm_query", + "choice_c": "qa_maths", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8718 + }, + { + "path": "audio-1501692468.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "alarm_set", + "choice_b": "transport_traffic", + "choice_c": "qa_currency", + "choice_d": "recommendation_locations", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8719 + }, + { + "path": "audio-1498563804.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "radio", + "choice_b": "play_radio", + "choice_c": "hue_lightdim", + "choice_d": "recommendation_events", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8720 + }, + { + "path": "audio-1501687487-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "calendar_set", + "choice_b": "play_music", + "choice_c": "datetime_query", + "choice_d": "iot_hue_lightup", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8721 + }, + { + "path": "audio-1502891234.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "play_audiobook", + "choice_b": "transport_taxi", + "choice_c": "general_joke", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8722 + }, + { + "path": "audio-1500039778-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "iot_wemo_on", + "choice_b": "lists_remove", + "choice_c": "calendar_query", + "choice_d": "play_music", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8723 + }, + { + "path": "audio-1500035458.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "general_quirky", + "choice_b": "remove", + "choice_c": "lists_query", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8724 + }, + { + "path": "audio-1502195883.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "weather_query", + "choice_b": "iot_wemo_off", + "choice_c": "qa_factoid", + "choice_d": "joke", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8725 + }, + { + "path": "audio-1488634162.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "qa_stock", + "choice_b": "qa_factoid", + "choice_c": "iot_hue_lighton", + "choice_d": "music_query", + "answer_gt": "iot_hue_lighton", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8726 + }, + { + "path": "audio-1498572644-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "post", + "choice_b": "qa_factoid", + "choice_c": "locations", + "choice_d": "datetime_convert", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8727 + }, + { + "path": "audio-1502199674.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "general_joke", + "choice_b": "music", + "choice_c": "datetime_query", + "choice_d": "takeaway_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8728 + }, + { + "path": "audio-1501778161.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "music_dislikeness", + "choice_b": "iot_cleaning", + "choice_c": "remove", + "choice_d": "sendemail", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8729 + }, + { + "path": "audio-1495369641-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "email_query", + "choice_b": "qa_maths", + "choice_c": "weather_query", + "choice_d": "iot_cleaning", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8730 + }, + { + "path": "audio-1501777288-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "email_query", + "choice_b": "audio_volume_up", + "choice_c": "locations", + "choice_d": "post", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8731 + }, + { + "path": "audio-1500979433-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "lists_createoradd", + "choice_b": "qa_currency", + "choice_c": "transport_taxi", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8732 + }, + { + "path": "audio-1502200860-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "iot_hue_lightup", + "choice_b": "cooking_recipe", + "choice_c": "audio_volume_mute", + "choice_d": "transport_query", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8733 + }, + { + "path": "audio-1434538601-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "cooking_recipe", + "choice_b": "play_music", + "choice_c": "weather_query", + "choice_d": "audio_volume_up", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8734 + }, + { + "path": "audio-1501753182.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "iot_coffee", + "choice_b": "recommendation_events", + "choice_c": "transport_query", + "choice_d": "alarm_query", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8735 + }, + { + "path": "audio-1502377117.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "alarm_set", + "choice_b": "recommendation_locations", + "choice_c": "play_game", + "choice_d": "datetime_query", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8736 + }, + { + "path": "audio-1488989926.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "qa_maths", + "choice_b": "music_query", + "choice_c": "alarm_set", + "choice_d": "transport_taxi", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8737 + }, + { + "path": "audio-1497883761-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "music_query", + "choice_b": "audio_volume_down", + "choice_c": "email_query", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8738 + }, + { + "path": "audio-1497885636.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "calendar_set", + "choice_b": "datetime_query", + "choice_c": "takeaway_order", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8739 + }, + { + "path": "audio-1497267569.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "social_post", + "choice_b": "post", + "choice_c": "calendar_query", + "choice_d": "alarm_set", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8740 + }, + { + "path": "audio-1488622599.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "iot_hue_lightchange", + "choice_b": "audio_volume_up", + "choice_c": "lists_createoradd", + "choice_d": "qa_factoid", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8741 + }, + { + "path": "audio-1488622268.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "cooking_recipe", + "choice_b": "datetime_query", + "choice_c": "play_radio", + "choice_d": "query", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8742 + }, + { + "path": "audio-1501689016-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "social_query", + "choice_b": "lists_query", + "choice_c": "iot_hue_lightdim", + "choice_d": "createoradd", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8743 + }, + { + "path": "audio-1497884047-headset.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "calendar_set", + "choice_b": "email_sendemail", + "choice_c": "takeaway_query", + "choice_d": "iot_hue_lightdim", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8744 + }, + { + "path": "audio-1498565844-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "qa_stock", + "choice_b": "audio_volume_down", + "choice_c": "calendar_query", + "choice_d": "qa_currency", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8745 + }, + { + "path": "audio-1494608083-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "social_post", + "choice_b": "joke", + "choice_c": "music_likeness", + "choice_d": "iot_hue_lightup", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8746 + }, + { + "path": "audio-1499691637-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "music_likeness", + "choice_b": "hue_lightdim", + "choice_c": "music_dislikeness", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8747 + }, + { + "path": "audio--1504197602.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "play_podcasts", + "choice_b": "news_query", + "choice_c": "radio", + "choice_d": "alarm_remove", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8748 + }, + { + "path": "audio-1498485590-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "takeaway_order", + "choice_b": "alarm_query", + "choice_c": "iot_wemo_on", + "choice_d": "qa_stock", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8749 + }, + { + "path": "audio--1504197770.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "audio_volume_up", + "choice_b": "play_audiobook", + "choice_c": "sendemail", + "choice_d": "iot_hue_lightup", + "answer_gt": "audio_volume_up", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8750 + }, + { + "path": "audio-1496265704-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "email_sendemail", + "choice_b": "music_query", + "choice_c": "takeaway_order", + "choice_d": "social_query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8751 + }, + { + "path": "audio-1495734126.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "news_query", + "choice_b": "recommendation_events", + "choice_c": "createoradd", + "choice_d": "alarm_remove", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8752 + }, + { + "path": "audio-1497557826-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "music_query", + "choice_b": "general_greet", + "choice_c": "audio_volume_down", + "choice_d": "news_query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8753 + }, + { + "path": "audio-1502309525-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_traffic", + "choice_b": "audio_volume_up", + "choice_c": "transport_ticket", + "choice_d": "play_music", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8754 + }, + { + "path": "audio-1502375629-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "lists_query", + "choice_b": "cooking_recipe", + "choice_c": "iot_cleaning", + "choice_d": "transport_taxi", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8755 + }, + { + "path": "audio-1490008615-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "play_audiobook", + "choice_b": "takeaway_query", + "choice_c": "transport_query", + "choice_d": "play_game", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8756 + }, + { + "path": "audio-1501766782-headset.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "email_query", + "choice_b": "transport_taxi", + "choice_c": "alarm_set", + "choice_d": "calendar_remove", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8757 + }, + { + "path": "audio-1490110858.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "cooking_recipe", + "choice_b": "takeaway_query", + "choice_c": "iot_coffee", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8758 + }, + { + "path": "audio-1434539865-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "calendar_remove", + "choice_b": "datetime_query", + "choice_c": "qa_currency", + "choice_d": "recommendation_movies", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8759 + }, + { + "path": "audio-1497866816-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "music_query", + "choice_b": "general_quirky", + "choice_c": "news_query", + "choice_d": "iot_hue_lightchange", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8760 + }, + { + "path": "audio-1501768606-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "lists_query", + "choice_b": "iot_hue_lighton", + "choice_c": "calendar_remove", + "choice_d": "iot_wemo_off", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8761 + }, + { + "path": "audio-1434533222.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "lists_remove", + "choice_b": "transport_traffic", + "choice_c": "audio_volume_other", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8762 + }, + { + "path": "audio-1434532510-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "audio_volume_down", + "choice_b": "alarm_query", + "choice_c": "lists_createoradd", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8763 + }, + { + "path": "audio-1490371121-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "qa_maths", + "choice_c": "calendar_set", + "choice_d": "iot_hue_lightoff", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8764 + }, + { + "path": "audio-1490368654-headset.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "recommendation_locations", + "choice_b": "weather_query", + "choice_c": "email_querycontact", + "choice_d": "email_query", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8765 + }, + { + "path": "audio-1497437606.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "iot_hue_lighton", + "choice_b": "audio_volume_other", + "choice_c": "createoradd", + "choice_d": "qa_definition", + "answer_gt": "iot_hue_lighton", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8766 + }, + { + "path": "audio-1434536750-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "iot_hue_lightoff", + "choice_b": "sendemail", + "choice_c": "email_sendemail", + "choice_d": "recommendation_movies", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8767 + }, + { + "path": "audio-1434541805-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "iot_hue_lightup", + "choice_b": "lists_remove", + "choice_c": "music_query", + "choice_d": "iot_hue_lightoff", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8768 + }, + { + "path": "audio-1490108335-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "news_query", + "choice_b": "qa_stock", + "choice_c": "iot_wemo_on", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8769 + }, + { + "path": "audio-1490261433.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "play_audiobook", + "choice_b": "general_joke", + "choice_c": "music_likeness", + "choice_d": "qa_maths", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8770 + }, + { + "path": "audio-1501777713-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "datetime_query", + "choice_b": "iot_hue_lightdim", + "choice_c": "news_query", + "choice_d": "iot_coffee", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8771 + }, + { + "path": "audio-1498565131-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "qa_maths", + "choice_b": "social_post", + "choice_c": "news_query", + "choice_d": "iot_hue_lightchange", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8772 + }, + { + "path": "audio--1506078372.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "weather_query", + "choice_b": "iot_coffee", + "choice_c": "email_addcontact", + "choice_d": "remove", + "answer_gt": "email_addcontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8773 + }, + { + "path": "audio-1499697278.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "alarm_remove", + "choice_b": "transport_ticket", + "choice_c": "alarm_set", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8774 + }, + { + "path": "audio-1488985555.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "general_quirky", + "choice_b": "general_greet", + "choice_c": "play_radio", + "choice_d": "iot_hue_lightup", + "answer_gt": "iot_hue_lightup", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8775 + }, + { + "path": "audio--1504195024-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "play_audiobook", + "choice_b": "radio", + "choice_c": "alarm_query", + "choice_d": "email_addcontact", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8776 + }, + { + "path": "audio-1497881502-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "general_quirky", + "choice_b": "qa_currency", + "choice_c": "iot_wemo_on", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8777 + }, + { + "path": "audio-1498482825.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "transport_taxi", + "choice_b": "audio_volume_mute", + "choice_c": "calendar_set", + "choice_d": "lists_remove", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8778 + }, + { + "path": "audio-1497558346-headset.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "music_settings", + "choice_b": "lists_createoradd", + "choice_c": "datetime_query", + "choice_d": "iot_hue_lightchange", + "answer_gt": "music_settings", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8779 + }, + { + "path": "audio-1497557669.flac", + "question": "How do you comprehend the speaker's intention from this audio?", + "choice_a": "lists_createoradd", + "choice_b": "cooking_recipe", + "choice_c": "createoradd", + "choice_d": "iot_hue_lightoff", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8780 + }, + { + "path": "audio--1504191460-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "remove", + "choice_b": "alarm_query", + "choice_c": "news_query", + "choice_d": "iot_wemo_off", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8781 + }, + { + "path": "audio-1498758469-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "play_podcasts", + "choice_b": "general_greet", + "choice_c": "alarm_set", + "choice_d": "iot_wemo_off", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8782 + }, + { + "path": "audio-1495369180.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "play_podcasts", + "choice_b": "hue_lightdim", + "choice_c": "transport_ticket", + "choice_d": "iot_hue_lightoff", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8783 + }, + { + "path": "audio-1489672057-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "play_music", + "choice_b": "post", + "choice_c": "audio_volume_other", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8784 + }, + { + "path": "audio-1501755900.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "alarm_query", + "choice_b": "general_greet", + "choice_c": "transport_query", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8785 + }, + { + "path": "audio-1502894611.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "remove", + "choice_b": "email_addcontact", + "choice_c": "datetime_query", + "choice_d": "qa_definition", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8786 + }, + { + "path": "audio-1501759308.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "lists_query", + "choice_b": "remove", + "choice_c": "email_query", + "choice_d": "calendar_remove", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8787 + }, + { + "path": "audio-1497620237-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "post", + "choice_b": "general_quirky", + "choice_c": "calendar_set", + "choice_d": "qa_factoid", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8788 + }, + { + "path": "audio-1502377499.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "cooking_recipe", + "choice_b": "play_podcasts", + "choice_c": "transport_ticket", + "choice_d": "music_likeness", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8789 + }, + { + "path": "audio--1504198820.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "iot_hue_lightchange", + "choice_b": "email_sendemail", + "choice_c": "cooking_recipe", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8790 + }, + { + "path": "audio-1502216570-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "calendar_query", + "choice_b": "qa_currency", + "choice_c": "general_quirky", + "choice_d": "radio", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8791 + }, + { + "path": "audio-1497192107-headset.flac", + "question": "infer the speaker's purpose from this audio.", + "choice_a": "play_radio", + "choice_b": "audio_volume_mute", + "choice_c": "takeaway_order", + "choice_d": "play_music", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8792 + }, + { + "path": "audio-1497872431.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "audio_volume_mute", + "choice_b": "calendar_query", + "choice_c": "iot_wemo_on", + "choice_d": "locations", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8793 + }, + { + "path": "audio-1502217514-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "play_music", + "choice_b": "remove", + "choice_c": "lists_query", + "choice_d": "hue_lightdim", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8794 + }, + { + "path": "audio-1498570575-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "datetime_query", + "choice_b": "query", + "choice_c": "alarm_remove", + "choice_d": "qa_definition", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8795 + }, + { + "path": "audio-1497270428-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "music", + "choice_b": "general_quirky", + "choice_c": "play_music", + "choice_d": "general_joke", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8796 + }, + { + "path": "audio-1498567418-headset.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "general_quirky", + "choice_b": "sendemail", + "choice_c": "email_addcontact", + "choice_d": "music_settings", + "answer_gt": "music_settings", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8797 + }, + { + "path": "audio-1494512806-headset.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "iot_wemo_on", + "choice_b": "lists_createoradd", + "choice_c": "qa_maths", + "choice_d": "play_radio", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8798 + }, + { + "path": "audio-1502896350.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "play_podcasts", + "choice_b": "recommendation_locations", + "choice_c": "lists_remove", + "choice_d": "qa_currency", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8799 + }, + { + "path": "audio-1502200532.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "takeaway_order", + "choice_b": "remove", + "choice_c": "play_audiobook", + "choice_d": "music_settings", + "answer_gt": "takeaway_order", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8800 + }, + { + "path": "audio-1502894736-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "iot_coffee", + "choice_b": "calendar_set", + "choice_c": "lists_remove", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8801 + }, + { + "path": "audio-1502199414.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "qa_maths", + "choice_b": "recommendation_locations", + "choice_c": "qa_stock", + "choice_d": "social_post", + "answer_gt": "qa_maths", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8802 + }, + { + "path": "audio-1494548477-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "remove", + "choice_b": "audio_volume_mute", + "choice_c": "news_query", + "choice_d": "recommendation_events", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8803 + }, + { + "path": "audio-1498646746-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "calendar_remove", + "choice_b": "joke", + "choice_c": "email_query", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8804 + }, + { + "path": "audio-1490104806-headset.flac", + "question": "What's your take on the speaker's objective in this recording?", + "choice_a": "lists_query", + "choice_b": "iot_coffee", + "choice_c": "iot_cleaning", + "choice_d": "email_addcontact", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8805 + }, + { + "path": "audio-1498564518-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "play_radio", + "choice_b": "play_audiobook", + "choice_c": "qa_definition", + "choice_d": "social_post", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8806 + }, + { + "path": "audio-1490368192.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "alarm_query", + "choice_b": "qa_definition", + "choice_c": "music_dislikeness", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8807 + }, + { + "path": "audio-1488622713.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "music_dislikeness", + "choice_b": "calendar_remove", + "choice_c": "general_quirky", + "choice_d": "play_music", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8808 + }, + { + "path": "audio-1502195883-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "email_sendemail", + "choice_b": "iot_wemo_off", + "choice_c": "social_query", + "choice_d": "query", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8809 + }, + { + "path": "audio-1492783811-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "iot_hue_lightdim", + "choice_b": "music", + "choice_c": "qa_definition", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8810 + }, + { + "path": "audio-1501608606.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "calendar_query", + "choice_b": "takeaway_order", + "choice_c": "social_post", + "choice_d": "general_joke", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8811 + }, + { + "path": "audio-1490356180-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "audio_volume_other", + "choice_b": "post", + "choice_c": "play_music", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8812 + }, + { + "path": "audio-1502101072-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "qa_definition", + "choice_b": "play_music", + "choice_c": "iot_cleaning", + "choice_d": "play_radio", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8813 + }, + { + "path": "audio-1490183647.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "music_settings", + "choice_b": "general_joke", + "choice_c": "general_quirky", + "choice_d": "email_sendemail", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8814 + }, + { + "path": "audio-1495370495-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "locations", + "choice_b": "news_query", + "choice_c": "transport_ticket", + "choice_d": "qa_maths", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8815 + }, + { + "path": "audio-1502891193-headset.flac", + "question": "What's your insight into the speaker's purpose after listening?", + "choice_a": "news_query", + "choice_b": "iot_wemo_off", + "choice_c": "iot_hue_lightoff", + "choice_d": "general_quirky", + "answer_gt": "iot_wemo_off", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8816 + }, + { + "path": "audio-1490108553-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "play_game", + "choice_b": "calendar_query", + "choice_c": "alarm_set", + "choice_d": "qa_factoid", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8817 + }, + { + "path": "audio-1499267205.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "hue_lightoff", + "choice_b": "general_quirky", + "choice_c": "email_addcontact", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8818 + }, + { + "path": "audio-1497622596.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "news_query", + "choice_c": "music_query", + "choice_d": "audio_volume_up", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8819 + }, + { + "path": "audio-1501157904-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "iot_hue_lightchange", + "choice_b": "play_game", + "choice_c": "music_likeness", + "choice_d": "play_music", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8820 + }, + { + "path": "audio-1490106394-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_query", + "choice_b": "music_likeness", + "choice_c": "cooking_recipe", + "choice_d": "hue_lightdim", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8821 + }, + { + "path": "audio-1502199616.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "music", + "choice_b": "qa_stock", + "choice_c": "email_addcontact", + "choice_d": "audio_volume_up", + "answer_gt": "audio_volume_up", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8822 + }, + { + "path": "audio-1500036864.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "lists_query", + "choice_b": "music_likeness", + "choice_c": "general_joke", + "choice_d": "qa_stock", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8823 + }, + { + "path": "audio-1490108775-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "qa_maths", + "choice_b": "calendar_query", + "choice_c": "transport_taxi", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8824 + }, + { + "path": "audio-1490193801-headset.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "email_sendemail", + "choice_b": "cooking_recipe", + "choice_c": "play_podcasts", + "choice_d": "sendemail", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8825 + }, + { + "path": "audio-1501685829-headset.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "general_greet", + "choice_b": "alarm_set", + "choice_c": "news_query", + "choice_d": "recommendation_locations", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8826 + }, + { + "path": "audio-1501415172.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "transport_traffic", + "choice_b": "qa_definition", + "choice_c": "lists_query", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8827 + }, + { + "path": "audio-1497880827-headset.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "qa_definition", + "choice_b": "qa_maths", + "choice_c": "weather_query", + "choice_d": "email_querycontact", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8828 + }, + { + "path": "audio-1494513921-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "news_query", + "choice_b": "qa_stock", + "choice_c": "iot_hue_lightchange", + "choice_d": "iot_hue_lightdim", + "answer_gt": "qa_stock", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8829 + }, + { + "path": "audio-1501758792-headset.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "datetime_query", + "choice_b": "post", + "choice_c": "email_querycontact", + "choice_d": "radio", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8830 + }, + { + "path": "audio-1497269825.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_cleaning", + "choice_b": "audio_volume_up", + "choice_c": "transport_taxi", + "choice_d": "qa_factoid", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8831 + }, + { + "path": "audio-1501688361.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "datetime_query", + "choice_b": "alarm_remove", + "choice_c": "calendar_remove", + "choice_d": "social_query", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8832 + }, + { + "path": "audio-1434542371.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "hue_lightoff", + "choice_b": "qa_stock", + "choice_c": "recommendation_movies", + "choice_d": "calendar_set", + "answer_gt": "recommendation_movies", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8833 + }, + { + "path": "audio-1499696168.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "calendar_remove", + "choice_b": "takeaway_order", + "choice_c": "iot_hue_lightchange", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8834 + }, + { + "path": "audio--1506079343-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "social_query", + "choice_b": "weather_query", + "choice_c": "play_game", + "choice_d": "recommendation_locations", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8835 + }, + { + "path": "audio-1490194270-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "music", + "choice_b": "iot_cleaning", + "choice_c": "audio_volume_mute", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8836 + }, + { + "path": "audio-1502194923-headset.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "transport_taxi", + "choice_b": "post", + "choice_c": "general_joke", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8837 + }, + { + "path": "audio-1499089424-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "transport_taxi", + "choice_b": "hue_lightdim", + "choice_c": "news_query", + "choice_d": "transport_traffic", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8838 + }, + { + "path": "audio-1501151625.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "transport_traffic", + "choice_b": "takeaway_query", + "choice_c": "iot_hue_lightchange", + "choice_d": "social_query", + "answer_gt": "iot_hue_lightchange", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8839 + }, + { + "path": "audio-1490007930-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "iot_coffee", + "choice_b": "music_likeness", + "choice_c": "social_query", + "choice_d": "alarm_remove", + "answer_gt": "social_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8840 + }, + { + "path": "audio-1499691920-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "email_addcontact", + "choice_b": "qa_definition", + "choice_c": "calendar_query", + "choice_d": "music_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8841 + }, + { + "path": "audio--1504192153-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "query", + "choice_b": "email_sendemail", + "choice_c": "alarm_set", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8842 + }, + { + "path": "audio-1501692051.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "music_settings", + "choice_b": "play_radio", + "choice_c": "qa_currency", + "choice_d": "alarm_remove", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8843 + }, + { + "path": "audio-1501767668-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "audio_volume_mute", + "choice_b": "weather_query", + "choice_c": "recommendation_events", + "choice_d": "lists_remove", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8844 + }, + { + "path": "audio-1495732964-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "play_music", + "choice_b": "iot_wemo_on", + "choice_c": "createoradd", + "choice_d": "iot_hue_lighton", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8845 + }, + { + "path": "audio-1498567113-headset.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "play_radio", + "choice_b": "hue_lightoff", + "choice_c": "hue_lightdim", + "choice_d": "audio_volume_down", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8846 + }, + { + "path": "audio-1499089971.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "takeaway_query", + "choice_b": "play_podcasts", + "choice_c": "recommendation_movies", + "choice_d": "audio_volume_up", + "answer_gt": "play_podcasts", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8847 + }, + { + "path": "audio-1501687491.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "audio_volume_other", + "choice_b": "qa_currency", + "choice_c": "qa_factoid", + "choice_d": "iot_hue_lighton", + "answer_gt": "qa_currency", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8848 + }, + { + "path": "audio-1434532541-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "general_quirky", + "choice_b": "lists_remove", + "choice_c": "transport_ticket", + "choice_d": "alarm_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8849 + }, + { + "path": "audio-1490264552-headset.flac", + "question": "What's your perspective on the speaker's intended effect in this recording?", + "choice_a": "alarm_remove", + "choice_b": "iot_hue_lightoff", + "choice_c": "general_joke", + "choice_d": "transport_ticket", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8850 + }, + { + "path": "audio-1501692631.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "alarm_set", + "choice_b": "post", + "choice_c": "remove", + "choice_d": "iot_coffee", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8851 + }, + { + "path": "audio-1490194432.flac", + "question": "What's your estimation of the speaker's motive in this audio?", + "choice_a": "email_query", + "choice_b": "lists_remove", + "choice_c": "music_settings", + "choice_d": "takeaway_order", + "answer_gt": "lists_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8852 + }, + { + "path": "audio-1501686116-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "news_query", + "choice_b": "recommendation_movies", + "choice_c": "transport_taxi", + "choice_d": "iot_hue_lightdim", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8853 + }, + { + "path": "audio-1501407300.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "email_sendemail", + "choice_b": "calendar_set", + "choice_c": "hue_lightoff", + "choice_d": "post", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8854 + }, + { + "path": "audio-1490356160-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "takeaway_query", + "choice_b": "social_post", + "choice_c": "general_quirky", + "choice_d": "transport_query", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8855 + }, + { + "path": "audio-1492784992.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "social_post", + "choice_b": "audio_volume_down", + "choice_c": "calendar_set", + "choice_d": "recommendation_locations", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8856 + }, + { + "path": "audio-1495732798.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "hue_lightdim", + "choice_b": "music_dislikeness", + "choice_c": "iot_coffee", + "choice_d": "iot_hue_lightup", + "answer_gt": "music_dislikeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8857 + }, + { + "path": "audio--1504195533-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "audio_volume_up", + "choice_b": "qa_currency", + "choice_c": "cooking_recipe", + "choice_d": "alarm_remove", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8858 + }, + { + "path": "audio-1502113207-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "remove", + "choice_b": "calendar_query", + "choice_c": "qa_definition", + "choice_d": "iot_wemo_on", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8859 + }, + { + "path": "audio-1500389999-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "music_query", + "choice_b": "email_sendemail", + "choice_c": "createoradd", + "choice_d": "iot_coffee", + "answer_gt": "iot_coffee", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8860 + }, + { + "path": "audio-1501758877.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "qa_definition", + "choice_b": "alarm_set", + "choice_c": "email_querycontact", + "choice_d": "transport_traffic", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8861 + }, + { + "path": "audio-1502197873-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "weather_query", + "choice_b": "audio_volume_other", + "choice_c": "alarm_query", + "choice_d": "iot_wemo_off", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8862 + }, + { + "path": "audio--1504194828.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "lists_createoradd", + "choice_b": "calendar_remove", + "choice_c": "play_podcasts", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8863 + }, + { + "path": "audio-1500979012-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "music_likeness", + "choice_b": "transport_taxi", + "choice_c": "qa_currency", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8864 + }, + { + "path": "audio-1497271364.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "weather_query", + "choice_b": "takeaway_query", + "choice_c": "qa_definition", + "choice_d": "calendar_remove", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8865 + }, + { + "path": "audio-1490103572.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "iot_hue_lightoff", + "choice_b": "hue_lightdim", + "choice_c": "lists_createoradd", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8866 + }, + { + "path": "audio-1494608621-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "play_game", + "choice_b": "createoradd", + "choice_c": "recommendation_movies", + "choice_d": "play_radio", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8867 + }, + { + "path": "audio-1498566690.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "play_audiobook", + "choice_c": "qa_factoid", + "choice_d": "audio_volume_up", + "answer_gt": "iot_hue_lightdim", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8868 + }, + { + "path": "audio-1501753191.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "iot_hue_lightoff", + "choice_b": "datetime_query", + "choice_c": "iot_hue_lightdim", + "choice_d": "transport_taxi", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8869 + }, + { + "path": "audio-1488984819.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "news_query", + "choice_b": "radio", + "choice_c": "recommendation_events", + "choice_d": "qa_factoid", + "answer_gt": "recommendation_events", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8870 + }, + { + "path": "audio-1499088159.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "email_query", + "choice_c": "transport_taxi", + "choice_d": "social_post", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8871 + }, + { + "path": "audio-1490105718-headset.flac", + "question": "How do you view the speaker's agenda in this audio?", + "choice_a": "sendemail", + "choice_b": "qa_definition", + "choice_c": "lists_query", + "choice_d": "lists_remove", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8872 + }, + { + "path": "audio-1501764070.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "qa_maths", + "choice_b": "calendar_set", + "choice_c": "email_query", + "choice_d": "audio_volume_down", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8873 + }, + { + "path": "audio-1498484091-headset.flac", + "question": "What's your belief about the speaker's intention after hearing this clip?", + "choice_a": "lists_query", + "choice_b": "iot_cleaning", + "choice_c": "play_music", + "choice_d": "iot_hue_lightup", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8874 + }, + { + "path": "audio-1500040478-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "lists_remove", + "choice_b": "social_post", + "choice_c": "datetime_convert", + "choice_d": "news_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8875 + }, + { + "path": "audio-1501765013.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "datetime_query", + "choice_b": "qa_maths", + "choice_c": "qa_definition", + "choice_d": "play_audiobook", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8876 + }, + { + "path": "audio-1498571512.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "audio_volume_up", + "choice_b": "email_querycontact", + "choice_c": "hue_lightoff", + "choice_d": "music_query", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8877 + }, + { + "path": "audio-1502197492.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "news_query", + "choice_b": "locations", + "choice_c": "audio_volume_down", + "choice_d": "datetime_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8878 + }, + { + "path": "audio-1499690211-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "takeaway_query", + "choice_b": "transport_query", + "choice_c": "transport_traffic", + "choice_d": "music", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8879 + }, + { + "path": "audio-1488982791.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "lists_query", + "choice_b": "iot_cleaning", + "choice_c": "recommendation_events", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8880 + }, + { + "path": "audio-1499361652.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "qa_maths", + "choice_b": "email_sendemail", + "choice_c": "general_quirky", + "choice_d": "takeaway_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8881 + }, + { + "path": "audio-1498563641-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "transport_ticket", + "choice_b": "lists_createoradd", + "choice_c": "calendar_remove", + "choice_d": "lists_query", + "answer_gt": "lists_createoradd", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8882 + }, + { + "path": "audio-1490956090-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "lists_query", + "choice_b": "takeaway_order", + "choice_c": "hue_lightoff", + "choice_d": "calendar_remove", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8883 + }, + { + "path": "audio-1434539121.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "general_quirky", + "choice_b": "hue_lightoff", + "choice_c": "query", + "choice_d": "alarm_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8884 + }, + { + "path": "audio-1497438531-headset.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "general_quirky", + "choice_b": "datetime_query", + "choice_c": "locations", + "choice_d": "lists_remove", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8885 + }, + { + "path": "audio-1490195068.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "audio_volume_up", + "choice_b": "datetime_convert", + "choice_c": "transport_taxi", + "choice_d": "email_sendemail", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8886 + }, + { + "path": "audio-1499265908.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "weather_query", + "choice_b": "music_likeness", + "choice_c": "createoradd", + "choice_d": "recommendation_movies", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8887 + }, + { + "path": "audio-1501688151-headset.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "calendar_remove", + "choice_b": "transport_ticket", + "choice_c": "email_sendemail", + "choice_d": "email_querycontact", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8888 + }, + { + "path": "audio-1499244679-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "email_query", + "choice_b": "alarm_query", + "choice_c": "audio_volume_down", + "choice_d": "recommendation_events", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8889 + }, + { + "path": "audio--1504197931-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "recommendation_events", + "choice_b": "music_likeness", + "choice_c": "play_music", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8890 + }, + { + "path": "audio-1488983597.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "email_sendemail", + "choice_b": "play_music", + "choice_c": "iot_wemo_off", + "choice_d": "createoradd", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8891 + }, + { + "path": "audio-1497449685.flac", + "question": "How do you assess the speaker's goal from this audio?", + "choice_a": "datetime_query", + "choice_b": "email_sendemail", + "choice_c": "iot_wemo_off", + "choice_d": "play_podcasts", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8892 + }, + { + "path": "audio-1490201183-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "email_addcontact", + "choice_b": "datetime_query", + "choice_c": "sendemail", + "choice_d": "query", + "answer_gt": "query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8893 + }, + { + "path": "audio-1494421825-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "email_sendemail", + "choice_b": "radio", + "choice_c": "qa_maths", + "choice_d": "transport_query", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8894 + }, + { + "path": "audio-1502376594-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "general_joke", + "choice_b": "general_quirky", + "choice_c": "iot_hue_lightoff", + "choice_d": "music_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8895 + }, + { + "path": "audio-1502309334.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "alarm_set", + "choice_c": "email_sendemail", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8896 + }, + { + "path": "audio-1497557205.flac", + "question": "What's your conclusion about the speaker's aim after listening?", + "choice_a": "music_query", + "choice_b": "qa_maths", + "choice_c": "audio_volume_mute", + "choice_d": "play_game", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8897 + }, + { + "path": "audio-1502310211-headset.flac", + "question": "What's your reflection on the speaker's goal after experiencing this audio?", + "choice_a": "transport_taxi", + "choice_b": "qa_factoid", + "choice_c": "audio_volume_mute", + "choice_d": "iot_hue_lightdim", + "answer_gt": "transport_taxi", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8898 + }, + { + "path": "audio-1434544537.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "general_greet", + "choice_b": "transport_taxi", + "choice_c": "lists_createoradd", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8899 + }, + { + "path": "audio-1490967181.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "news_query", + "choice_b": "iot_hue_lighton", + "choice_c": "alarm_set", + "choice_d": "alarm_query", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8900 + }, + { + "path": "audio-1494513631-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "radio", + "choice_b": "datetime_query", + "choice_c": "general_quirky", + "choice_d": "alarm_remove", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8901 + }, + { + "path": "audio-1492785200.flac", + "question": "How do you read the speaker's intention in this sound clip?", + "choice_a": "transport_ticket", + "choice_b": "social_post", + "choice_c": "weather_query", + "choice_d": "hue_lightdim", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8902 + }, + { + "path": "audio-1490367667.flac", + "question": "What's your understanding of the speaker's intent in this audio?", + "choice_a": "audio_volume_mute", + "choice_b": "music_likeness", + "choice_c": "play_radio", + "choice_d": "general_joke", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8903 + }, + { + "path": "audio-1497866047.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "music_likeness", + "choice_b": "general_greet", + "choice_c": "lists_createoradd", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8904 + }, + { + "path": "audio-1498575851.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "music", + "choice_b": "recommendation_locations", + "choice_c": "general_quirky", + "choice_d": "email_sendemail", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8905 + }, + { + "path": "audio-1502194762-headset.flac", + "question": "What do you deduce the speaker's intention to be from this recording?", + "choice_a": "recommendation_movies", + "choice_b": "qa_stock", + "choice_c": "iot_cleaning", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8906 + }, + { + "path": "audio-1502891087.flac", + "question": "What's your assessment of the speaker's purpose in this audio?", + "choice_a": "post", + "choice_b": "email_sendemail", + "choice_c": "music_likeness", + "choice_d": "news_query", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8907 + }, + { + "path": "audio-1502890281.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "recommendation_events", + "choice_b": "play_podcasts", + "choice_c": "lists_createoradd", + "choice_d": "alarm_query", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8908 + }, + { + "path": "audio-1502114423-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "iot_hue_lightup", + "choice_b": "datetime_query", + "choice_c": "iot_coffee", + "choice_d": "play_radio", + "answer_gt": "play_radio", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8909 + }, + { + "path": "audio-1497872829.flac", + "question": "What do you think the speaker's end goal is in this audio?", + "choice_a": "createoradd", + "choice_b": "radio", + "choice_c": "music_settings", + "choice_d": "iot_hue_lightoff", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8910 + }, + { + "path": "audio--1505405792.flac", + "question": "What's your hypothesis about the speaker's intention in this audio?", + "choice_a": "music_query", + "choice_b": "play_audiobook", + "choice_c": "post", + "choice_d": "createoradd", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8911 + }, + { + "path": "audio-1502299420.flac", + "question": "What's your speculation on the speaker's aim in this recording?", + "choice_a": "email_sendemail", + "choice_b": "music_query", + "choice_c": "cooking_recipe", + "choice_d": "iot_hue_lightchange", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8912 + }, + { + "path": "audio-1490801084-headset.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "general_quirky", + "choice_b": "music_query", + "choice_c": "calendar_set", + "choice_d": "audio_volume_down", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8913 + }, + { + "path": "audio-1502895101-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "music_dislikeness", + "choice_b": "iot_wemo_off", + "choice_c": "lists_createoradd", + "choice_d": "play_audiobook", + "answer_gt": "play_audiobook", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8914 + }, + { + "path": "audio-1490368753-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "music_settings", + "choice_b": "transport_query", + "choice_c": "takeaway_query", + "choice_d": "transport_traffic", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8915 + }, + { + "path": "audio-1490369455.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "play_music", + "choice_b": "music_dislikeness", + "choice_c": "play_podcasts", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8916 + }, + { + "path": "audio-1501698268-headset.flac", + "question": "How do you see the speaker's intention in this audio clip?", + "choice_a": "play_music", + "choice_b": "calendar_query", + "choice_c": "general_joke", + "choice_d": "takeaway_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8917 + }, + { + "path": "audio-1495376058.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "play_audiobook", + "choice_b": "iot_hue_lightdim", + "choice_c": "qa_factoid", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8918 + }, + { + "path": "audio-1497868233-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "social_post", + "choice_b": "createoradd", + "choice_c": "general_joke", + "choice_d": "post", + "answer_gt": "social_post", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8919 + }, + { + "path": "audio-1501692764-headset.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "general_quirky", + "choice_b": "transport_traffic", + "choice_c": "query", + "choice_d": "social_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8920 + }, + { + "path": "audio-1490262120.flac", + "question": "What's your idea of the speaker's purpose in this sound recording?", + "choice_a": "social_query", + "choice_b": "sendemail", + "choice_c": "lists_createoradd", + "choice_d": "play_game", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8921 + }, + { + "path": "audio-1498485027-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "play_radio", + "choice_b": "recommendation_events", + "choice_c": "cooking_recipe", + "choice_d": "transport_traffic", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8922 + }, + { + "path": "audio-1489154192.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "transport_traffic", + "choice_b": "locations", + "choice_c": "music_likeness", + "choice_d": "play_music", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8923 + }, + { + "path": "audio-1498757508-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "iot_hue_lighton", + "choice_b": "email_query", + "choice_c": "email_addcontact", + "choice_d": "music_likeness", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8924 + }, + { + "path": "audio--1506520943.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "locations", + "choice_b": "music_query", + "choice_c": "datetime_convert", + "choice_d": "audio_volume_other", + "answer_gt": "music_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8925 + }, + { + "path": "audio-1501415244.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "email_query", + "choice_b": "iot_hue_lightoff", + "choice_c": "email_sendemail", + "choice_d": "lists_createoradd", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8926 + }, + { + "path": "audio-1501767734-headset.flac", + "question": "What's your guess at the speaker's objective after listening to this recording?", + "choice_a": "email_querycontact", + "choice_b": "play_music", + "choice_c": "email_query", + "choice_d": "audio_volume_mute", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8927 + }, + { + "path": "audio--1504192882-headset.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "general_joke", + "choice_b": "alarm_query", + "choice_c": "hue_lightoff", + "choice_d": "recommendation_movies", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8928 + }, + { + "path": "audio-1500389012-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "calendar_set", + "choice_b": "calendar_query", + "choice_c": "iot_wemo_off", + "choice_d": "takeaway_order", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8929 + }, + { + "path": "audio-1502101919.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "music_likeness", + "choice_b": "cooking_recipe", + "choice_c": "play_podcasts", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8930 + }, + { + "path": "audio-1502374642-headset.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "alarm_set", + "choice_b": "general_quirky", + "choice_c": "qa_currency", + "choice_d": "news_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8931 + }, + { + "path": "audio-1488622303.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "qa_stock", + "choice_b": "iot_hue_lightdim", + "choice_c": "lists_remove", + "choice_d": "cooking_recipe", + "answer_gt": "cooking_recipe", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8932 + }, + { + "path": "audio-1490200110-headset.flac", + "question": "What do you think is the speaker's intention after listening to this audio?", + "choice_a": "audio_volume_down", + "choice_b": "iot_wemo_off", + "choice_c": "alarm_remove", + "choice_d": "transport_ticket", + "answer_gt": "transport_ticket", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8933 + }, + { + "path": "audio--1505404351-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "alarm_set", + "choice_b": "createoradd", + "choice_c": "transport_taxi", + "choice_d": "qa_definition", + "answer_gt": "alarm_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8934 + }, + { + "path": "audio-1499089953.flac", + "question": "What do you think the speaker is attempting to convey in this audio?", + "choice_a": "alarm_query", + "choice_b": "locations", + "choice_c": "calendar_remove", + "choice_d": "datetime_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8935 + }, + { + "path": "audio-1497884638-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "email_sendemail", + "choice_b": "play_radio", + "choice_c": "audio_volume_up", + "choice_d": "iot_wemo_on", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8936 + }, + { + "path": "audio-1499088835-headset.flac", + "question": "What do you suppose is the speaker's motive in this recording?", + "choice_a": "general_greet", + "choice_b": "qa_definition", + "choice_c": "music_dislikeness", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8937 + }, + { + "path": "audio-1495377539-headset.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "weather_query", + "choice_b": "post", + "choice_c": "transport_query", + "choice_d": "play_game", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8938 + }, + { + "path": "audio-1490263507.flac", + "question": "What's your theory on the speaker's motive in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "iot_cleaning", + "choice_c": "general_joke", + "choice_d": "calendar_query", + "answer_gt": "datetime_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8939 + }, + { + "path": "audio-1499244208-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "calendar_remove", + "choice_b": "iot_hue_lightchange", + "choice_c": "lists_remove", + "choice_d": "recommendation_locations", + "answer_gt": "calendar_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8940 + }, + { + "path": "audio-1497881678.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "email_addcontact", + "choice_b": "iot_hue_lighton", + "choice_c": "createoradd", + "choice_d": "weather_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8941 + }, + { + "path": "audio-1502200593-headset.flac", + "question": "How do you understand the speaker's intended message in this clip?", + "choice_a": "qa_currency", + "choice_b": "weather_query", + "choice_c": "qa_stock", + "choice_d": "joke", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8942 + }, + { + "path": "audio-1496265649-headset.flac", + "question": "What's your sense of the speaker's intention from this audio?", + "choice_a": "qa_factoid", + "choice_b": "calendar_set", + "choice_c": "general_quirky", + "choice_d": "social_query", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8943 + }, + { + "path": "audio-1500418460-headset.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "transport_query", + "choice_b": "music_likeness", + "choice_c": "email_sendemail", + "choice_d": "recommendation_events", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8944 + }, + { + "path": "audio--1505404290.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "play_radio", + "choice_b": "music_query", + "choice_c": "takeaway_query", + "choice_d": "joke", + "answer_gt": "takeaway_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8945 + }, + { + "path": "audio-1490201741.flac", + "question": "What's your impression of the speaker's target message in this sound recording?", + "choice_a": "email_sendemail", + "choice_b": "calendar_query", + "choice_c": "recommendation_events", + "choice_d": "alarm_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8946 + }, + { + "path": "audio-1490371992-headset.flac", + "question": "What intention do you attribute to the speaker in this audio?", + "choice_a": "email_query", + "choice_b": "takeaway_order", + "choice_c": "calendar_remove", + "choice_d": "music_dislikeness", + "answer_gt": "email_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8947 + }, + { + "path": "audio-1498565956.flac", + "question": "What's your take on the speaker's desired response after hearing this clip?", + "choice_a": "transport_ticket", + "choice_b": "recommendation_movies", + "choice_c": "email_querycontact", + "choice_d": "lists_query", + "answer_gt": "lists_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8948 + }, + { + "path": "audio-1498568433-headset.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "alarm_query", + "choice_b": "audio_volume_other", + "choice_c": "datetime_query", + "choice_d": "play_radio", + "answer_gt": "alarm_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8949 + }, + { + "path": "audio-1490262120-headset.flac", + "question": "What's your interpretation of the speaker's reason for this recording?", + "choice_a": "cooking_recipe", + "choice_b": "alarm_remove", + "choice_c": "play_game", + "choice_d": "iot_coffee", + "answer_gt": "play_game", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8950 + }, + { + "path": "audio-1498646110.flac", + "question": "What's your perception of the speaker's goal after hearing this clip?", + "choice_a": "iot_hue_lightdim", + "choice_b": "iot_cleaning", + "choice_c": "iot_hue_lightoff", + "choice_d": "iot_hue_lightup", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8951 + }, + { + "path": "audio-1495372243.flac", + "question": "What do you think the speaker's plan is based on this recording?", + "choice_a": "calendar_query", + "choice_b": "play_music", + "choice_c": "email_query", + "choice_d": "qa_definition", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8952 + }, + { + "path": "audio-1488969797.flac", + "question": "What do you believe the speaker is trying to achieve with this message?", + "choice_a": "lists_query", + "choice_b": "weather_query", + "choice_c": "transport_ticket", + "choice_d": "news_query", + "answer_gt": "weather_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8953 + }, + { + "path": "audio-1489497388-headset.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "post", + "choice_b": "play_podcasts", + "choice_c": "locations", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8954 + }, + { + "path": "audio-1497029534.flac", + "question": "How do you perceive the speaker's agenda in this sound clip?", + "choice_a": "locations", + "choice_b": "play_game", + "choice_c": "general_quirky", + "choice_d": "email_addcontact", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8955 + }, + { + "path": "audio-1502218188-headset.flac", + "question": "What's your analysis of the speaker's intent in this audio?", + "choice_a": "email_sendemail", + "choice_b": "transport_taxi", + "choice_c": "qa_currency", + "choice_d": "cooking_recipe", + "answer_gt": "email_sendemail", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8956 + }, + { + "path": "audio-1499694002.flac", + "question": "How do you interpret the speaker's desired impact from this audio?", + "choice_a": "audio_volume_down", + "choice_b": "play_music", + "choice_c": "general_joke", + "choice_d": "iot_cleaning", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8957 + }, + { + "path": "audio-1502099312.flac", + "question": "What's your guess about the speaker's intention in this audio?", + "choice_a": "news_query", + "choice_b": "transport_query", + "choice_c": "datetime_query", + "choice_d": "email_querycontact", + "answer_gt": "news_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8958 + }, + { + "path": "audio-1501754331-headset.flac", + "question": "What's your take on what the speaker is trying to communicate in this recording?", + "choice_a": "iot_hue_lightdim", + "choice_b": "datetime_convert", + "choice_c": "play_music", + "choice_d": "calendar_query", + "answer_gt": "play_music", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8959 + }, + { + "path": "audio-1499693890-headset.flac", + "question": "How would you interpret the speaker's aim in this audio?", + "choice_a": "takeaway_order", + "choice_b": "email_sendemail", + "choice_c": "general_quirky", + "choice_d": "calendar_query", + "answer_gt": "calendar_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8960 + }, + { + "path": "audio--1505407196-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "social_post", + "choice_b": "alarm_remove", + "choice_c": "play_game", + "choice_d": "email_sendemail", + "answer_gt": "alarm_remove", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8961 + }, + { + "path": "audio-1502196155.flac", + "question": "How do you figure the speaker's intention from this audio?", + "choice_a": "social_post", + "choice_b": "social_query", + "choice_c": "calendar_query", + "choice_d": "general_quirky", + "answer_gt": "general_quirky", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8962 + }, + { + "path": "audio-1490365526-headset.flac", + "question": "After listening, what do you think the speaker's desired outcome is?", + "choice_a": "alarm_remove", + "choice_b": "iot_hue_lightoff", + "choice_c": "transport_taxi", + "choice_d": "calendar_set", + "answer_gt": "calendar_set", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8963 + }, + { + "path": "audio-1497025499.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "alarm_set", + "choice_b": "general_joke", + "choice_c": "takeaway_order", + "choice_d": "iot_hue_lightdim", + "answer_gt": "general_joke", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8964 + }, + { + "path": "audio-1490706169.flac", + "question": "What's your conjecture about the speaker's purpose in this sound clip?", + "choice_a": "transport_query", + "choice_b": "iot_hue_lightup", + "choice_c": "qa_maths", + "choice_d": "weather_query", + "answer_gt": "transport_query", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8965 + }, + { + "path": "audio-1502376487-headset.flac", + "question": "What's your understanding of what the speaker intends with this audio?", + "choice_a": "iot_wemo_off", + "choice_b": "transport_traffic", + "choice_c": "news_query", + "choice_d": "qa_factoid", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8966 + }, + { + "path": "audio-1501609043.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "iot_hue_lightoff", + "choice_b": "iot_wemo_on", + "choice_c": "alarm_set", + "choice_d": "play_audiobook", + "answer_gt": "iot_hue_lightoff", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8967 + }, + { + "path": "audio-1434539611.flac", + "question": "What do you think the speaker's underlying message is in this audio?", + "choice_a": "transport_taxi", + "choice_b": "qa_factoid", + "choice_c": "play_audiobook", + "choice_d": "iot_wemo_on", + "answer_gt": "qa_factoid", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8968 + }, + { + "path": "audio-1490288564-headset.flac", + "question": "What's your view on the speaker's objective after hearing this recording?", + "choice_a": "play_radio", + "choice_b": "transport_traffic", + "choice_c": "hue_lightoff", + "choice_d": "iot_cleaning", + "answer_gt": "iot_cleaning", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8969 + }, + { + "path": "audio-1502309483.flac", + "question": "What's your opinion on the speaker's goal in this sound clip?", + "choice_a": "datetime_query", + "choice_b": "alarm_remove", + "choice_c": "alarm_set", + "choice_d": "email_querycontact", + "answer_gt": "email_querycontact", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8970 + }, + { + "path": "audio-1501151476-headset.flac", + "question": "What do you think the speaker's message is intended to be in this audio?", + "choice_a": "sendemail", + "choice_b": "qa_definition", + "choice_c": "joke", + "choice_d": "datetime_convert", + "answer_gt": "qa_definition", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8971 + }, + { + "path": "audio-1501774395-headset.flac", + "question": "How do you judge the speaker's aim in this sound clip?", + "choice_a": "music_likeness", + "choice_b": "calendar_query", + "choice_c": "iot_wemo_off", + "choice_d": "cooking_recipe", + "answer_gt": "music_likeness", + "task_name": "Speaker_Intent_Classification", + "dataset_name": "slurp", + "uniq_id": 8972 + }, + { + "path": "U0UNvoRitTg_00023-lLmMZZrFGCw_00022.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8973 + }, + { + "path": "yVWutBwaJME_00020-JVm6b7QviH8_00010.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8974 + }, + { + "path": "f7Cym6TZfAA_00004-f7Cym6TZfAA_00005.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8975 + }, + { + "path": "cG4FJ7wp2z4_00012-doK4sIjaoTA_00016.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8976 + }, + { + "path": "m5-lh8k3Y4M_00006-_uVms69aRHY_00001.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8977 + }, + { + "path": "PfQFgX_pBfU_00002-1R3XTTBorKk_00022.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8978 + }, + { + "path": "SUq9njwBwn0_00002-FUiiruCPVTI_00002.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8979 + }, + { + "path": "Jb_HWf5_2E0_00007-vhzF7dhGJzI_00003.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8980 + }, + { + "path": "Xlvi4H6dKxg_00001-8Us5MqS0QNk_00002.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8981 + }, + { + "path": "71DxnZPSimA_00003-XM4tc46aJu8_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8982 + }, + { + "path": "xMVOmZULqTk_00001-zMFEL2tfeGA_00002.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8983 + }, + { + "path": "qv0yN-6L4wA_00003-ihBB7_GYohM_00003.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8984 + }, + { + "path": "Gzam5nAQQgc_00003-D3bwdXeY87M_00003.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8985 + }, + { + "path": "iTefLRE0Y_o_00002-EbGsEyE5UXc_00006.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8986 + }, + { + "path": "pBFbAzUYb90_00008-z2dCpFO1U8A_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8987 + }, + { + "path": "gNnNpOn_dV0_00094-5ablueV_1tw_00007.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8988 + }, + { + "path": "S_Mr5l8yRiU_00002-BxlF0-hQOcI_00005.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8989 + }, + { + "path": "xCMXM3aIpPQ_00005-SYK-37GSJWM_00014.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8990 + }, + { + "path": "DNsQx_QmigQ_00004-nDpaVYtKQUo_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8991 + }, + { + "path": "Qwaw5FLYnr8_00001-AqrOntSB71E_00003.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8992 + }, + { + "path": "2SQgxXSsh68_00024-Bs1ucqT6C_w_00015.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8993 + }, + { + "path": "kJ87LdPguuM_00004-aIiBNSrPmhY_00060.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8994 + }, + { + "path": "koRHZBBT_VI_00002-A1vEHrrVdFg_00004.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8995 + }, + { + "path": "leGeIt4bPHs_00012-4dPZAlWNv1I_00010.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8996 + }, + { + "path": "KHVJaN5lrq4_00006-SJz7TqbRv7M_00008.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8997 + }, + { + "path": "3FcrpSSVgEU_00002-CksSx4W0QZo_00019.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8998 + }, + { + "path": "5KEClGiyvMQ_00002-je-E0NIk6cM_00010.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 8999 + }, + { + "path": "yeAfGSnnGOM_00014-hDgolbSElUM_00018.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9000 + }, + { + "path": "HvBwJrc_-ns_00009-qo94xxqllQ0_00011.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9001 + }, + { + "path": "lBlTm5DS5VQ_00012-p7ZGqJ1K3rk_00007.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9002 + }, + { + "path": "D1Gl7M8p6pg_00007-pQaCSml7NW4_00023.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9003 + }, + { + "path": "JRYO8cIrJMc_00002-x6O8O_n5YZ8_00018.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9004 + }, + { + "path": "E50E5hbggyM_00003-CFgJCAw60hs_00028.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9005 + }, + { + "path": "8rYu90wXbaE_00003-GXVt48XOV-s_00001.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9006 + }, + { + "path": "LiuMxzOZ374_00010-jlYnGTpkcSE_00001.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9007 + }, + { + "path": "wMPjKZ4vQec_00001-cshhP-eEdm0_00004.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9008 + }, + { + "path": "Rkq0LOe94GY_00009-hisw7lDJQp4_00007.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9009 + }, + { + "path": "zz-4OJGmRmg_00014-EYZaUHNcx_8_00006.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9010 + }, + { + "path": "KZIY_Rpi0ug_00004-nggtIVFcD-Q_00003.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9011 + }, + { + "path": "OLoCQMhFNrI_00002-CbF8AurFLR8_00013.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9012 + }, + { + "path": "zjuonbHlU2Q_00004-RHpr11_ZQ3I_00014.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9013 + }, + { + "path": "QUC5ptsIu7k_00004-y35AZ0aP0hU_00014.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9014 + }, + { + "path": "n9neWqTdAIM_00004-JQtDfEz08aU_00023.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9015 + }, + { + "path": "mMsAKsN9XBk_00001-nBXbTtDDC1Y_00004.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9016 + }, + { + "path": "nwv-Y58mse8_00002-xc0tqY_-7_I_00003.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9017 + }, + { + "path": "SVy18zoTenQ_00005-OhFEclsyhwk_00012.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9018 + }, + { + "path": "9TVyLvhBmbo_00010-KSsKOIYLKU8_00001.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9019 + }, + { + "path": "BMgByP9ZCck_00001-jLJfU-yNxtw_00001.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9020 + }, + { + "path": "oKpLEjAsWEA_00002-TMCCVk1tTnE_00004.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9021 + }, + { + "path": "FtIkMMVq3bs_00002-AWDlX_Pmmqw_00004.wav", + "question": "The speech features how many speakers?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9022 + }, + { + "path": "iyHOYmDAYio_00004-ZvoXeFwyMH8_00007.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9023 + }, + { + "path": "LtVaUs2nBBs_00002-w59g68aiVR0_00001.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9024 + }, + { + "path": "85tdkVMF2y0_00005-UM8t5TAmU4c_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9025 + }, + { + "path": "1D22Tgx1k4A_00033-dV8Q5-tEB6A_00001.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9026 + }, + { + "path": "_cWY5jrhnrc_00015-2nf1qU4Kfsc_00020.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9027 + }, + { + "path": "urvlAaiea64_00002-rurIC-bn1xw_00001.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9028 + }, + { + "path": "fz4MYyBgFpQ_00008-xwEWIYfouFw_00004.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9029 + }, + { + "path": "CqeHGHlScSI_00002-zSHf5VEQVyE_00003.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9030 + }, + { + "path": "F8kg0LEuKCE_00003-VJ__L3391hU_00015.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9031 + }, + { + "path": "8Bto4IColtU_00010-7jKLn1sTpYY_00005.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9032 + }, + { + "path": "cYOxRhtUFQY_00024-cFQBQaqngsc_00015.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9033 + }, + { + "path": "hjkh3-GTMLw_00006-8AOqndLyKdg_00006.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9034 + }, + { + "path": "bAAgkmE2p6Q_00001-wWmt0eUteSg_00002.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9035 + }, + { + "path": "5EuCD6gyU5M_00020-32B9pRIxqJI_00003.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9036 + }, + { + "path": "2ekOREjUP4k_00001-ZgtAh6BFRIk_00012.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9037 + }, + { + "path": "XRHe5XV3ENk_00001-cKYTZp8y3Ao_00016.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9038 + }, + { + "path": "Yocizg57t48_00003-4bwnchJ3GdQ_00008.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9039 + }, + { + "path": "je-E0NIk6cM_00005-9jsnPvz4stg_00089.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9040 + }, + { + "path": "J9WK4j7mZCA_00031-UKQ_6-OyPnw_00001.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9041 + }, + { + "path": "ObPcP4a-a4I_00008-zcypXF7OVKo_00011.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9042 + }, + { + "path": "0a4dF2TtdOM_00002-xAAgE4KpR8g_00001.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9043 + }, + { + "path": "UbFEQgMHnfQ_00029-dePmiDpodMA_00009.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9044 + }, + { + "path": "WeVHst8swP0_00020-aMfPWRLKR0g_00015.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9045 + }, + { + "path": "umcTq_Pv2B8_00002-_IoMoOPdGmY_00003.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9046 + }, + { + "path": "HnUcElU-IxE_00005-9xPiK-K2DNo_00008.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9047 + }, + { + "path": "P-7wcXaLUUs_00017-zDxGuLm7Z6I_00006.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9048 + }, + { + "path": "9VR7wckrdP8_00002-dNdCzyv4VQE_00013.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9049 + }, + { + "path": "R_tlrhgP52Q_00003-JABvGlNIK6w_00001.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9050 + }, + { + "path": "L-Z9OI-WwoY_00001-hyUfye8vNI8_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9051 + }, + { + "path": "1y9XR62zKvs_00002-mXBZPFu7i5E_00002.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9052 + }, + { + "path": "ONwb6-2w2sA_00008-GcPsUzXov4M_00001.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9053 + }, + { + "path": "tFc4EciiEPU_00002-bd4yG4JRIUU_00010.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9054 + }, + { + "path": "rAqtoL86w2Q_00007-YffYxcL0PQk_00004.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9055 + }, + { + "path": "_h10Pr6m2A0_00004-Rb8QqCqgCEM_00007.wav", + "question": "How many people are delivering this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9056 + }, + { + "path": "8cfyJEV7hP8_00026-RlXVj2MlnnE_00005.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9057 + }, + { + "path": "y3_iiY6v4W0_00003-whMJs28Tu_s_00018.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9058 + }, + { + "path": "Vv0Zm9kisrU_00006-BOb190-piZY_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9059 + }, + { + "path": "ETDT0kqfjfA_00028-zDkgVesX7NU_00001.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9060 + }, + { + "path": "WXqRMEeAMQ4_00006-UNYe2joO2IE_00003.wav", + "question": "How many people are delivering this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9061 + }, + { + "path": "w9bqihhUvCo_00014-g6KZST3wats_00002.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9062 + }, + { + "path": "rsO79xUsxlU_00043-dmils9KmvK4_00005.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9063 + }, + { + "path": "HtBbOQ6Yab8_00029-OC_EISBqooQ_00003.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9064 + }, + { + "path": "y0UFutwJ-ow_00054-008ZEVovmQo_00002.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9065 + }, + { + "path": "PMk0szsm0dE_00003-cc_U4QV8aco_00002.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9066 + }, + { + "path": "qSObOdJjRnA_00036-oU8-tDn-9Y0_00007.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9067 + }, + { + "path": "108kzMff1Jw_00008-L9J3ZProgGw_00011.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9068 + }, + { + "path": "_gqH11CZpRs_00004-1uqrS4yUk34_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9069 + }, + { + "path": "sKkwzUqXEXs_00002-r7qsqKmv1L0_00003.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9070 + }, + { + "path": "0txtXuZBqC4_00074-BEOISoSqXqU_00005.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9071 + }, + { + "path": "WDo_jz_PwNQ_00001-_7Pf9G_wkEY_00001.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9072 + }, + { + "path": "MdTcExMSIZI_00001-5bE7zSyjNSs_00089.wav", + "question": "The speech captures how many different voices?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9073 + }, + { + "path": "s-kH45WFljQ_00008-6KXf1QT9GC0_00034.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9074 + }, + { + "path": "d4yJly3db4U_00001-bGMCGEtTArs_00001.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9075 + }, + { + "path": "XRfJw9GmoeM_00004-P8O80I9OgR4_00003.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9076 + }, + { + "path": "4IAzvK87HqI_00008-RhtEKbRiTnU_00013.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9077 + }, + { + "path": "WhGVQSUE1-Q_00001-aBH7pmHztTQ_00003.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9078 + }, + { + "path": "YwqXV22SA4U_00002-YwqXV22SA4U_00003.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9079 + }, + { + "path": "CgtVM7uD-Ws_00004-Vchu9G6SDuI_00005.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9080 + }, + { + "path": "dinctLMTXmM_00011-N6Xz7-IZvVo_00003.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9081 + }, + { + "path": "Qxm16zbou2g_00007-GpSk6dLbkGM_00010.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9082 + }, + { + "path": "JjEzHhHsdss_00002-h36oGrTJbfU_00002.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9083 + }, + { + "path": "Eoo7tBT2gQU_00001-GJK5UH-knVI_00002.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9084 + }, + { + "path": "tZT90vCdKGk_00003-4tlCKgb3LvU_00016.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9085 + }, + { + "path": "mk1UmZlF7XU_00021-BETpEZOK66c_00002.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9086 + }, + { + "path": "WNFrOOBHlVY_00002-CK0Y8FNi6r8_00008.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9087 + }, + { + "path": "eSGPnNa31kU_00006-7CGOm2DVh_U_00014.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9088 + }, + { + "path": "cNvo7HjL8ys_00015-WJPpL58RruA_00008.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9089 + }, + { + "path": "2LSiFyZN5-Y_00020-bwYAk5HVPCk_00002.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9090 + }, + { + "path": "Jkx4BMLN4kw_00002-J8OJUhBI_ak_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9091 + }, + { + "path": "bi5GkeonHnY_00001-XNErZsk5h0g_00003.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9092 + }, + { + "path": "G3QojmXtMM4_00029-mYoEm9nvntA_00004.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9093 + }, + { + "path": "U9duwXJyt0k_00008-3i95mBs9Zgo_00001.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9094 + }, + { + "path": "IURLbUMafXs_00006-vFSuB1VE8zw_00006.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9095 + }, + { + "path": "03SSllwNkGk_00003-BAq0dOMcMbY_00009.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9096 + }, + { + "path": "ywNiQmOo0lQ_00006-kBj2lWjDT3s_00003.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9097 + }, + { + "path": "MGswAhcv2Gw_00001-cEBM_wQStZw_00014.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9098 + }, + { + "path": "o3WNLAhGN2g_00006-dLhPmQVbcaw_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9099 + }, + { + "path": "OS1I7a4LNww_00005-Klv3JuHv-jM_00008.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9100 + }, + { + "path": "Oa-gvLYpOec_00017-YEhj0NORvBM_00006.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9101 + }, + { + "path": "KahonI0gjuA_00001-mGVjlgIVNFI_00003.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9102 + }, + { + "path": "t-8vN3uC-CU_00004-ftqN24vX-t0_00008.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9103 + }, + { + "path": "bU-9ZOW0lMU_00001-DjkYU8gcEvM_00012.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9104 + }, + { + "path": "2iL0P9T7pYY_00004-Q3lLoioyzyY_00001.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9105 + }, + { + "path": "ya6VNZp-pXw_00018-gm6PJowclv0_00023.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9106 + }, + { + "path": "NWsXYEX4olc_00008-awK0SDaltDw_00003.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9107 + }, + { + "path": "yJ5DFQQbOZU_00011-TFIZ9vWg6EE_00007.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9108 + }, + { + "path": "cMI5F_qxd4M_00003-98qWWArcL3Y_00002.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9109 + }, + { + "path": "zWE8NG5oSOA_00016-O77p-6-AqP4_00057.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9110 + }, + { + "path": "s3ZWWI1GJZ8_00005-R4S2c-NiPfA_00007.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9111 + }, + { + "path": "UJFb8jDbWl0_00005-vT9LdyTsG2M_00001.wav", + "question": "How many people are delivering this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9112 + }, + { + "path": "SRAQpJx6cWQ_00001-qo94xxqllQ0_00009.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9113 + }, + { + "path": "Rq3OWTy0nD0_00002-0kxar61XyMc_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9114 + }, + { + "path": "9rPVtEBBjFs_00005-oUHGKXT4Aew_00018.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9115 + }, + { + "path": "EyHxjyBaAGw_00016-ZOFJhtebvuU_00007.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9116 + }, + { + "path": "2awUa_hS7DY_00008-x5fvCg-HGdo_00062.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9117 + }, + { + "path": "YEVZp443BBQ_00008-Y2lVYYS0afY_00012.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9118 + }, + { + "path": "rWaBVelzaFk_00020-CYb_opUpiw4_00013.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9119 + }, + { + "path": "ROV--v_ZoDE_00024-IXrZP_6JUdA_00012.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9120 + }, + { + "path": "OHDRhzSfICk_00037-OHDRhzSfICk_00006.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9121 + }, + { + "path": "5yNbNqsQ6JM_00006-8CM7ZiZtRaE_00003.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9122 + }, + { + "path": "na8-QEFmj44_00003-Ud8gVoDVDU0_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9123 + }, + { + "path": "QanuGhOhb9A_00009-fO-gHrA3Bfk_00008.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9124 + }, + { + "path": "wENZk-gv4Pg_00001-rVtEZqVtEwg_00004.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9125 + }, + { + "path": "Du0qgBHW5Mg_00002-x5TwkglqxG4_00003.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9126 + }, + { + "path": "Qq_ncaVFAMs_00014-8RP7uetAVM8_00001.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9127 + }, + { + "path": "avOhqSI8AbU_00006-fT3lv_iqN28_00001.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9128 + }, + { + "path": "DLQY19QumUc_00032-S1BLPmNvLek_00008.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9129 + }, + { + "path": "cyr2RlVuDVY_00001-Qv7RVdd9LtI_00033.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9130 + }, + { + "path": "mGPj8-GkZkU_00002-xmUNXr-sRaA_00006.wav", + "question": "How many people are delivering this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9131 + }, + { + "path": "Sn3685sVfq0_00001-O77p-6-AqP4_00071.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9132 + }, + { + "path": "XQQoRntKO8k_00002-k3DGwIlknw8_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9133 + }, + { + "path": "YDKFrw_i5Io_00013-7VW6w_Uz_kg_00001.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9134 + }, + { + "path": "lFEBeMFuL_M_00010-6SBGrr4iFp0_00038.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9135 + }, + { + "path": "xH3Pp_5yxOk_00013-G3QojmXtMM4_00010.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9136 + }, + { + "path": "lFHQNZGlgng_00007-oLYkxi95OEA_00002.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9137 + }, + { + "path": "kvqNCn2dtw0_00004-eZfqINabWik_00015.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9138 + }, + { + "path": "v1kzk7zfEuc_00001-PUW0_GO7UX8_00007.wav", + "question": "How many people are delivering this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9139 + }, + { + "path": "D5JvrPaUCJw_00003-t4gpfyjRkqQ_00007.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9140 + }, + { + "path": "VDJi9gg6RhI_00008-k2mGVGCILIk_00003.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9141 + }, + { + "path": "8ycnZc6cHC8_00003-DpDEYfw0VlY_00002.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9142 + }, + { + "path": "G31c_RqjNiQ_00005-Vod1r77aj8E_00021.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9143 + }, + { + "path": "2LSiFyZN5-Y_00013-_DvLV9aicU4_00011.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9144 + }, + { + "path": "Fejq2JzedIQ_00016-YvRKvdS43t8_00010.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9145 + }, + { + "path": "ntgkeQhp-0c_00021-ntgkeQhp-0c_00009.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9146 + }, + { + "path": "QeDtLS5lj2s_00003-v1aPRoiBZ6E_00002.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9147 + }, + { + "path": "eHb5mpgcHsE_00014-2iZQFpRit3E_00006.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9148 + }, + { + "path": "Fq9xHBNoPtA_00008-fsXwDEEnj10_00005.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9149 + }, + { + "path": "Bq8dqnt1HdQ_00001-4ngJFREfZVI_00009.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9150 + }, + { + "path": "2wZLZSmQgSk_00001-jBqtwUYODlU_00018.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9151 + }, + { + "path": "G9VW2MPvX40_00021-OODAmigQ1Z4_00004.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9152 + }, + { + "path": "ugBD8-FlBh0_00002-5JTgeBLrkUc_00003.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9153 + }, + { + "path": "uqJOS2DkgmM_00001-04859am4Z8k_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9154 + }, + { + "path": "C5f1GK7hD-c_00030-YhV39sDmDYA_00007.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9155 + }, + { + "path": "IOL7cbC-i18_00009-DSwXeR4NfwA_00002.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9156 + }, + { + "path": "G515om3QPyA_00001-MEa183Zyf5g_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9157 + }, + { + "path": "EBRN94RlDAk_00016-KxtgAuy1rmA_00002.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9158 + }, + { + "path": "1ne9RX6tN5A_00008-20aJUK-V4rg_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9159 + }, + { + "path": "xs7wQ9wxtZ0_00008-3wHTyqxyDqk_00009.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9160 + }, + { + "path": "Q-2Au8iHje0_00001-TdhfxAIWANo_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9161 + }, + { + "path": "N_524p1QzXQ_00005-N_524p1QzXQ_00008.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9162 + }, + { + "path": "b8n1oh-9UB4_00001-Xlvi4H6dKxg_00001.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9163 + }, + { + "path": "Wt-Hc3-PdgI_00019-Ao-AU5v682Y_00004.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9164 + }, + { + "path": "7byARrJfhlI_00008-QfELI4G0Fro_00010.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9165 + }, + { + "path": "cfRB26fxHaw_00002-cfRB26fxHaw_00017.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9166 + }, + { + "path": "IeRkCV_1jmk_00003-7u-rskIpVxE_00005.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9167 + }, + { + "path": "4Mjv5tVqYUU_00002-4Mjv5tVqYUU_00002.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9168 + }, + { + "path": "NIVVbWBeOPU_00013-1MNbrgPasLA_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9169 + }, + { + "path": "0NQZ9mnI5XQ_00008-HNHLYfX4qhM_00003.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9170 + }, + { + "path": "LlrzVUoR9rw_00001-NxBndFq3G_w_00002.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9171 + }, + { + "path": "GGpq6aWa0zA_00001-5CHLE5D7NaM_00007.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9172 + }, + { + "path": "EBVYy1H_eTM_00008-4wcoqK5SSWo_00001.wav", + "question": "How many people are delivering this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9173 + }, + { + "path": "WySIhmZZ_MA_00009-swAilGrlGnU_00001.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9174 + }, + { + "path": "ja2o1wEznow_00003-bZELSKogKJM_00010.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9175 + }, + { + "path": "JVJXcI4Zudk_00057-JVJXcI4Zudk_00065.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9176 + }, + { + "path": "iJVf4nfI9lQ_00002-N26rCBM4NZ4_00004.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9177 + }, + { + "path": "oQSA_tR628s_00003-lWTwjQnsVV4_00001.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9178 + }, + { + "path": "0-OTCt3gm_s_00015-dzTAszYtckM_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9179 + }, + { + "path": "iCq6rrXVcfk_00005-mDv_dDjO9n4_00003.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9180 + }, + { + "path": "LxEGHW6Lbu8_00004-SY_FpBvyoKE_00013.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9181 + }, + { + "path": "Ifz1XY9Ob3c_00017-XbkMcvnNq3g_00002.wav", + "question": "The speech features how many speakers?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9182 + }, + { + "path": "eIgqgaMrRQ0_00002-TEcDrsqajOA_00017.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9183 + }, + { + "path": "h0P7XPHskpE_00001-PMHMpcHg_YM_00019.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9184 + }, + { + "path": "PD9_9m2dqiA_00002-V6AXVM19MtU_00010.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9185 + }, + { + "path": "Xqq9mAyYyMo_00015-ZOsjLV0XXlQ_00003.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9186 + }, + { + "path": "VmpQfMU8bHQ_00001-z5POm7W7vOc_00012.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9187 + }, + { + "path": "mdaA-DjnOfg_00002-plGeZ2hDSlY_00010.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9188 + }, + { + "path": "4iLNk3yaJgc_00028-HpHOcifSuXM_00003.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9189 + }, + { + "path": "GYwSSRJ9DSM_00001-4mqW_eLvfDI_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9190 + }, + { + "path": "AgaDSqjx9Gc_00007-qrQSXBh9v_0_00003.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9191 + }, + { + "path": "bZELSKogKJM_00007-EiPQ4whOiuk_00006.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9192 + }, + { + "path": "UQtyjaWG4O4_00002-suSnxN1Wm8c_00002.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9193 + }, + { + "path": "NuokWhpheSk_00007-fSoU9Uhv7UA_00014.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9194 + }, + { + "path": "Q-O0MtBvSgw_00007-iJVf4nfI9lQ_00005.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9195 + }, + { + "path": "_j2RydnWAU8_00007-5PNDuB9nW-0_00018.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9196 + }, + { + "path": "Ns7ocpRhDD8_00001-3n6hd9gxBQ4_00005.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9197 + }, + { + "path": "jRJhagwY92s_00014-LrBefvUxHsQ_00027.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9198 + }, + { + "path": "Yh4FbOwv0NU_00001-UkO3JubUo8Y_00013.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9199 + }, + { + "path": "KebqWBNGQH0_00003-NeUucCm4q-o_00004.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9200 + }, + { + "path": "dn1JNgHzyZg_00006-hS4iRzg0T2M_00010.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9201 + }, + { + "path": "4-FPKEV3qZY_00001-dvtP66g_mYE_00002.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9202 + }, + { + "path": "3FESCkcOQqQ_00002-YXXrbu2zdG0_00006.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9203 + }, + { + "path": "Jamhv-uq4JI_00009-bA3gUpox2qU_00001.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9204 + }, + { + "path": "dYV8y111GBI_00003-dzLfm4YxN9M_00009.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9205 + }, + { + "path": "9Rg_q0gWSJE_00007-VOoZngq6GFA_00002.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9206 + }, + { + "path": "DCMWL-rM_jY_00026-6sdguNXxtBc_00008.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9207 + }, + { + "path": "kunZO5i2JnA_00003-lIHPlKJvozo_00002.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9208 + }, + { + "path": "JEdQ7DL-RXw_00002-jZK-29B50g0_00009.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9209 + }, + { + "path": "xT7i2cpJI9w_00086-kFVA2rUj2LY_00004.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9210 + }, + { + "path": "TKQ2eU8u-TE_00027-Gq5HrgnqlK8_00003.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9211 + }, + { + "path": "mJo37SEoITU_00007-zSc9H971RRo_00001.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9212 + }, + { + "path": "fcaDDXnhKW0_00011-RkZJhsyqak8_00010.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9213 + }, + { + "path": "np_JZN_eo0E_00001-5cu92kGSGCY_00027.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9214 + }, + { + "path": "01q-YLBhP6I_00006-fJ2rsQ3N60c_00004.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9215 + }, + { + "path": "CA27ZHZ56Ts_00005-YYl7evawZnw_00004.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9216 + }, + { + "path": "82u0tUvOb6c_00040-377w7_cw5U0_00013.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9217 + }, + { + "path": "yV8NNCEDlek_00070-MYfw_v7Cqcg_00031.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9218 + }, + { + "path": "0q9ZMc67NhY_00002-IIsSUemzSFY_00011.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9219 + }, + { + "path": "1Mq97vK5fPI_00001-xC6jYNIKKxY_00001.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9220 + }, + { + "path": "987UrtpOII8_00003-hKcg4JN09fw_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9221 + }, + { + "path": "yBAUUw9ovKw_00006-G3w1yj7Nhmw_00004.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9222 + }, + { + "path": "s4QQAwcUNYI_00001-zs6LYGoctZ4_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9223 + }, + { + "path": "m0pirbsgB-c_00025-0W0BRZvSCTU_00007.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9224 + }, + { + "path": "P_V-TnaJJXQ_00005-ZvoXeFwyMH8_00016.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9225 + }, + { + "path": "qtkJNZsjRRA_00002-mva1pG9I4EU_00001.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9226 + }, + { + "path": "ggm8jEhV8xs_00004-uEW56YuJXk0_00004.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9227 + }, + { + "path": "Nicl3Q05JQ0_00009-FOhECbBrehg_00002.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9228 + }, + { + "path": "q-qvheefvFw_00002-Mpr9wqUuLQA_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9229 + }, + { + "path": "4iLNk3yaJgc_00029-IRuhsRDUjOI_00008.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9230 + }, + { + "path": "HQ1AFT2c0JE_00004-iCC_z72DRmc_00007.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9231 + }, + { + "path": "ToOr53u4u_A_00004-kqE_hyI6Ago_00004.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9232 + }, + { + "path": "spPCNcMKaj4_00007-yjcw5ySDvUY_00020.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9233 + }, + { + "path": "VFyPhJmxazI_00033-pV6cJibX6Ik_00001.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9234 + }, + { + "path": "2iV9HT3E1Ls_00007-5xGJYiNH2Jw_00009.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9235 + }, + { + "path": "Mc7XFTXvcNA_00001-vtCEvIVhIKE_00014.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9236 + }, + { + "path": "XcZRqtFEis0_00007-IXjBDMJc5tE_00023.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9237 + }, + { + "path": "BRo6B3N-hcY_00002-duM9ky6Ce4Y_00001.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9238 + }, + { + "path": "r_gK4yvLnOI_00001-XsiRQ8XiWOk_00003.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9239 + }, + { + "path": "EI0BcP32ins_00044-sIH47kQ0VJg_00005.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9240 + }, + { + "path": "_eDcERY4p-A_00005-Mm326dbpKYQ_00014.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9241 + }, + { + "path": "EX6GfxU7vGA_00004-TxahHq341DY_00003.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9242 + }, + { + "path": "eXJkcQT3dsQ_00003-TSkYk-YtaK8_00008.wav", + "question": "The speech captures how many different voices?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9243 + }, + { + "path": "xk8okzebW5c_00005-XBxLJGCPjFM_00056.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9244 + }, + { + "path": "EtdEnAZ_6UA_00001-s4Jpzb2oOKg_00005.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9245 + }, + { + "path": "bC6tUNIR9V0_00063-1hEr7qKRKL4_00007.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9246 + }, + { + "path": "QqO6kjUmbR8_00001-llAe9CEv6S0_00001.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9247 + }, + { + "path": "EmtHk7Iflp8_00002-FfKvI_HhRmM_00004.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9248 + }, + { + "path": "8DIzctOYPgQ_00004-rRCEPhMS8XQ_00001.wav", + "question": "The speech captures how many different voices?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9249 + }, + { + "path": "RNOrbahjFtk_00012-4-A3igAu_aQ_00002.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9250 + }, + { + "path": "6kGwhudccdg_00004-tlgg5yrEKsE_00003.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9251 + }, + { + "path": "p08Vaf307-A_00023-PbT967mHtDE_00010.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9252 + }, + { + "path": "L0civBljdW8_00001-PCClICprMwc_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9253 + }, + { + "path": "EHIPNo4WDrQ_00005-5Ciy1E1gYRc_00091.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9254 + }, + { + "path": "M2k_enZdWu4_00039-aaebCc11h2I_00005.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9255 + }, + { + "path": "vhzF7dhGJzI_00002-5W5ODrDdRns_00009.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9256 + }, + { + "path": "P_Um2VdIKmA_00076-4iLNk3yaJgc_00010.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9257 + }, + { + "path": "Vr3xG-5f9VI_00017-Vr3xG-5f9VI_00021.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9258 + }, + { + "path": "z-7C6WRL46c_00081-eP9bCmTlPlU_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9259 + }, + { + "path": "5G_ldPtfJQY_00010-jtmH2J2aYYA_00004.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9260 + }, + { + "path": "Xi8YB5xzExY_00009-m5-lh8k3Y4M_00010.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9261 + }, + { + "path": "tWpZt8roR3E_00043-q8SWx5H2tCA_00003.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9262 + }, + { + "path": "Qv7RVdd9LtI_00015-qZZJOcyyFOo_00006.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9263 + }, + { + "path": "jI2mcN6F6SA_00005-6Fw_sEav5TI_00003.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9264 + }, + { + "path": "u1iaCkCVTg0_00003-6kGwhudccdg_00005.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9265 + }, + { + "path": "msSyAnuw614_00001-IFEP003wMM0_00001.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9266 + }, + { + "path": "lfrJYPLyYwM_00034-3AMtpxgdeEs_00002.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9267 + }, + { + "path": "ftuQ0LL5HHo_00003-jsg4tSMumcw_00004.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9268 + }, + { + "path": "wMEajDh4fck_00017-kL3rndZCdvM_00017.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9269 + }, + { + "path": "lFEBeMFuL_M_00110-h1szbjS__2w_00014.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9270 + }, + { + "path": "h1szbjS__2w_00048-MtDAcXNZhQk_00007.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9271 + }, + { + "path": "6MR4tIkDTpg_00005-7J4N-z-ptUk_00006.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9272 + }, + { + "path": "y2N2MRJdT1Y_00010-WzmT9KmMVnY_00007.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9273 + }, + { + "path": "U8xqFgZutJ0_00003-3OERtMNQ0Is_00001.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9274 + }, + { + "path": "2Pq16ZKsZro_00014-RMDXpJ5e_kE_00052.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9275 + }, + { + "path": "CRcLZM0GnRs_00014-Bem9gqpcg6w_00005.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9276 + }, + { + "path": "0VpwYRLTKwM_00022-NnZ8irurCUU_00010.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9277 + }, + { + "path": "EvCyt2keqW4_00006-MSiccNrf9o4_00001.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9278 + }, + { + "path": "S1-TGXJX5gA_00003-q_8hyO29wWk_00003.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9279 + }, + { + "path": "LY0qvP1PcBI_00004-WzmT9KmMVnY_00011.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9280 + }, + { + "path": "AXjkhN1FXfI_00002-oxw1P5kbaAE_00001.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9281 + }, + { + "path": "xKWYozNgnRM_00001-wMiWdqfQvA0_00015.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9282 + }, + { + "path": "FPTxi3r2Sio_00039-FPTxi3r2Sio_00039.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9283 + }, + { + "path": "0IOynLx6Av8_00023-ssCWAalFYYY_00016.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9284 + }, + { + "path": "6CW-VQGYtUs_00006-VCsiuFxyrSU_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9285 + }, + { + "path": "xdHZvAB6Ioc_00020-SrpBcop7aFU_00008.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9286 + }, + { + "path": "dZDFZaR4CeU_00005-97jYhz7tmcg_00004.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9287 + }, + { + "path": "VncqAJt4iFU_00008-jLhbxKoivQU_00006.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9288 + }, + { + "path": "11xowx4MBO4_00002-UG6uJmM8FaY_00003.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9289 + }, + { + "path": "1CiBWgsHT4E_00001-MYfw_v7Cqcg_00015.wav", + "question": "How many people are delivering this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9290 + }, + { + "path": "e0kcnIzAMl4_00001-cVOqgcS_eRQ_00014.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9291 + }, + { + "path": "XsiRQ8XiWOk_00001-sQIqfA-I_Ew_00005.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9292 + }, + { + "path": "c_aEjTGGhZ4_00002-e211QU3kkb4_00007.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9293 + }, + { + "path": "NKJp3rNrMnY_00030-sKLVuGUGZKs_00003.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9294 + }, + { + "path": "8Fqh5NjkR-g_00008-8Fqh5NjkR-g_00013.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9295 + }, + { + "path": "WDG5r-GMZas_00004-vy8sQ82o0fM_00001.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9296 + }, + { + "path": "7jKLn1sTpYY_00007-NF7MIl6A3ug_00004.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9297 + }, + { + "path": "bRyxvGe10gA_00002-XsUUakkoA9E_00001.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9298 + }, + { + "path": "yVfXlncBQB4_00007-mO7wl7C1QZI_00005.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9299 + }, + { + "path": "91MtjruZrdU_00015-9wcs8O4I9u8_00001.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9300 + }, + { + "path": "DOsLMpphQ_M_00028-xw9f4hGyXyQ_00010.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9301 + }, + { + "path": "Dm3Htyp-iI4_00002-qQCowHad-G4_00001.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9302 + }, + { + "path": "1OJyP3MFPjc_00073-1OJyP3MFPjc_00062.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9303 + }, + { + "path": "InGU1HXJ6J0_00001-0yv4tCKP3Hs_00003.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9304 + }, + { + "path": "_M-AnFs7E_4_00002-rT9_GBVNQ30_00005.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9305 + }, + { + "path": "5dltpwMAG5s_00001-7EdoXiE3e1Q_00005.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9306 + }, + { + "path": "vpZkTc57vZA_00005-Iji-zD3DTa8_00001.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9307 + }, + { + "path": "XrTdAX_KO7Y_00002-nRLipPY4Z08_00001.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9308 + }, + { + "path": "KCDxXirtBoI_00011-moAaXexhWAE_00001.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9309 + }, + { + "path": "SQ1QX3A5Ruc_00001-C5WNk_2IC6Y_00006.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9310 + }, + { + "path": "i27cuuD6G5I_00026-ObvIB1teaa0_00014.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9311 + }, + { + "path": "MQJpzDsBPfc_00003-CFO1gnqx9Rc_00003.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9312 + }, + { + "path": "xNnoRBZWeEQ_00001-8J1CnIq9uIo_00028.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9313 + }, + { + "path": "4g1mKWebt3c_00001-iTw7EKjgeaI_00005.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9314 + }, + { + "path": "tSWJWrsAC_M_00005-sSU4jVcbHU4_00007.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9315 + }, + { + "path": "lFEBeMFuL_M_00022-tswc1vx5chY_00008.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9316 + }, + { + "path": "ryhLNl7xKG8_00009-yel3Y1KqTZ8_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9317 + }, + { + "path": "ac_9D_OpFJI_00001-9DVX_4LZ-D4_00006.wav", + "question": "The speech features how many speakers?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9318 + }, + { + "path": "Kq89N5Vb4Rk_00001-evuhgZfrwh8_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9319 + }, + { + "path": "JURnlnD6smA_00001-DcAhTUKXx9I_00007.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9320 + }, + { + "path": "YSHJrM6PFFE_00003-NOYKT8Jz3Tk_00007.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9321 + }, + { + "path": "iVBzgO-b5z0_00009-Bn-SNuseJiM_00001.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9322 + }, + { + "path": "nm9ge3yQaZE_00015-0L6pigD_dGg_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9323 + }, + { + "path": "6jXazEYi3P8_00001-LQQ1COemsOM_00002.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9324 + }, + { + "path": "jyTYRG0Hb4A_00005-DTQU1gsTcVY_00001.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9325 + }, + { + "path": "TBlcvADDi1c_00007-GkALww0NKTc_00002.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9326 + }, + { + "path": "sUeAVe5v7k4_00009-Pw7K0Sd2Qp4_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9327 + }, + { + "path": "2aq0MvfsfTs_00017-CzGkMEe7AyI_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9328 + }, + { + "path": "kXv2ev_gPkI_00002-x-JQVDl2EsU_00003.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9329 + }, + { + "path": "oKmR9wj2Xgo_00009-wImrFMaIsLA_00002.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9330 + }, + { + "path": "VCF5Hxe1F9Y_00014-oGsHYgqao2g_00001.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9331 + }, + { + "path": "dv2VMH0YM10_00009-hjlw5bHcbrQ_00002.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9332 + }, + { + "path": "8llUxZ3zaxY_00004-BMFMHGvKdgE_00002.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9333 + }, + { + "path": "w3oO-uHpn9s_00003-R6dg-JoNu_E_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9334 + }, + { + "path": "Z4M-Z0AYee8_00052-cY3-UkuZhJM_00007.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9335 + }, + { + "path": "bUO2y99OgPk_00002-kzkURZvzF70_00003.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9336 + }, + { + "path": "UTRUsPfPN2c_00001-PruP9LfY1fA_00001.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9337 + }, + { + "path": "CMvQastGhf0_00001-Ik3ySw5_OvM_00006.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9338 + }, + { + "path": "6qmmQ93AtI8_00002-SzhH5O1VtQE_00002.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9339 + }, + { + "path": "_FqC1kkr5Dw_00001-_FqC1kkr5Dw_00006.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9340 + }, + { + "path": "Vy9-9JeqcpM_00011-oNwR9SfVoqk_00002.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9341 + }, + { + "path": "LkBrGkkK64Q_00005-YmaD-eqxi7k_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9342 + }, + { + "path": "5bE7zSyjNSs_00032-1qDlSrPwX5s_00001.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9343 + }, + { + "path": "CFEDwCw8xr0_00006-dEpOs4y0-9Y_00003.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9344 + }, + { + "path": "J9lHsKG98U8_00007-utrA-v8pPm4_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9345 + }, + { + "path": "GtAHVLPqHLc_00012-jBcao1QfsYQ_00006.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9346 + }, + { + "path": "Pl79vE7tbUA_00008-bHXB9_VE9jk_00010.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9347 + }, + { + "path": "j96kxslsnyI_00011-h1DMAOVZ88o_00002.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9348 + }, + { + "path": "vGtaoU9d9BE_00005-nKMMXPaJ1Y0_00001.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9349 + }, + { + "path": "OmZATN7CrqY_00001-XenT4dFmggg_00004.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9350 + }, + { + "path": "tqdtbBSd4HI_00050-EtYu1IT8RzY_00001.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9351 + }, + { + "path": "y0UFutwJ-ow_00038-Adpb9Dn0PoM_00050.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9352 + }, + { + "path": "JAHXrKcVZTM_00012-GnHTN8Ob5pk_00004.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9353 + }, + { + "path": "XsJZ_SuFgR4_00001-HM6dEZ3NxAs_00001.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9354 + }, + { + "path": "IwB3vu418E4_00006-dCpykP6iJnA_00002.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9355 + }, + { + "path": "7CVeukwpMoM_00006-Ff4GN89IpTU_00005.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9356 + }, + { + "path": "m0pirbsgB-c_00073-91WBENcf6_o_00004.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9357 + }, + { + "path": "xKWYozNgnRM_00002-qi17JX-gBjc_00011.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9358 + }, + { + "path": "sezhmCve6as_00004-fGhSkJg4Sd4_00006.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9359 + }, + { + "path": "CgCvXNssXpQ_00002-fIXU5ExruI0_00005.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9360 + }, + { + "path": "ZNT2uhs3jF4_00001-c_aEjTGGhZ4_00003.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9361 + }, + { + "path": "n6j-UuSijgc_00006-VCEs8oAQAiM_00019.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9362 + }, + { + "path": "dYK2s3YM_o8_00002-lBIEfndWRI8_00001.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9363 + }, + { + "path": "zRLjy2nL_4w_00001-VTwVoIapW-Y_00015.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9364 + }, + { + "path": "j96kxslsnyI_00006-vJp5AfMhqhw_00003.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9365 + }, + { + "path": "Ep1gG3R68o4_00010-xfiZv0N3wvY_00010.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9366 + }, + { + "path": "4qBx8uXR1I4_00016-P0SOuWxBE-I_00004.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9367 + }, + { + "path": "d-SooYXE3Fc_00010-Beg96w0fPL8_00003.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9368 + }, + { + "path": "WcLzSLJ8RB8_00050-6sdguNXxtBc_00016.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9369 + }, + { + "path": "rsdueh8VUDk_00004-R12JMJ6ZcXo_00003.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9370 + }, + { + "path": "hcTds_fNQOQ_00002-UJFb8jDbWl0_00004.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9371 + }, + { + "path": "69XxH60bd4Y_00005-cyV9KCZP634_00014.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9372 + }, + { + "path": "a_X9yNacnoo_00001-UG6uJmM8FaY_00013.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9373 + }, + { + "path": "8e-XM55GGsk_00012-TMCCVk1tTnE_00004.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9374 + }, + { + "path": "pdi8PFlsM2A_00002-VITHKnnDF6o_00018.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9375 + }, + { + "path": "qMtrlaIokFA_00007-wjYZksl-CIo_00011.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9376 + }, + { + "path": "HzTwS9_Mm_o_00001-94Emo9rceak_00006.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9377 + }, + { + "path": "2Er-cV85fc4_00017-K2_D_tFdAgY_00022.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9378 + }, + { + "path": "xyuuDGFQdCM_00007-HYix__TaweM_00006.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9379 + }, + { + "path": "VSrjCY1Ro-g_00011-9T-RW9V_q4U_00001.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9380 + }, + { + "path": "oaKsZSu4sUc_00010-hV3oY_uKTLg_00004.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9381 + }, + { + "path": "g7SptlsGbaA_00008-KtST9n5smJA_00004.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9382 + }, + { + "path": "FX_m2by6-uc_00002-FX_m2by6-uc_00012.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9383 + }, + { + "path": "1A1AZ-l7P1c_00001-EWVd8qVp4RM_00001.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9384 + }, + { + "path": "YIEBxZxaUVM_00005-1jvx-ItXTN8_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9385 + }, + { + "path": "b4wl7nEZ2wE_00007-eM-Y97d-3XU_00003.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9386 + }, + { + "path": "z-7C6WRL46c_00044-itQzkwHEA6A_00001.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9387 + }, + { + "path": "NCFG73oAvUE_00004-3MLhKvfsiOQ_00009.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9388 + }, + { + "path": "8721zEnsLyA_00005-L8aRiQFMfQo_00004.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9389 + }, + { + "path": "CzQPAaPrC-E_00026-n_OWWyfyyo4_00001.wav", + "question": "The speech captures how many different voices?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9390 + }, + { + "path": "gNnNpOn_dV0_00025-uFcO1HTrRoM_00007.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9391 + }, + { + "path": "4uDtI4bTDbQ_00006-vTPw9CM_Iv8_00001.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9392 + }, + { + "path": "n3JBoirGzlc_00007-Os0CeHDvITE_00001.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9393 + }, + { + "path": "eTg_ONDWG10_00002-XrKM_LrA6_U_00008.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9394 + }, + { + "path": "U8xqFgZutJ0_00005-whfc-H_HSgQ_00007.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9395 + }, + { + "path": "HwJqhjP1T_4_00005-w7CfqmaKgn0_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9396 + }, + { + "path": "w9bqihhUvCo_00009-fCRpdMJGKJM_00021.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9397 + }, + { + "path": "kiV_l66ceaE_00013-sV0I6HzTZGI_00016.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9398 + }, + { + "path": "iTi65KCm8b0_00002-iC3frQUkwfw_00011.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9399 + }, + { + "path": "9TNhTMDAGuE_00003-wtC4hhd4ZoI_00001.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9400 + }, + { + "path": "qtsNbxgPngA_00003-7Gd4Fz2UNxQ_00032.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9401 + }, + { + "path": "RMDXpJ5e_kE_00111-ik0addgjsVM_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9402 + }, + { + "path": "A7PqUWo-4og_00005-d8a75TYQUVc_00007.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9403 + }, + { + "path": "NNT3V-dgnyA_00011-70z-Im2hzlg_00001.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9404 + }, + { + "path": "qpEPc2KJNjc_00003-p4X5vKRsRME_00004.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9405 + }, + { + "path": "Y4HrnOAeDt4_00003-AEwXlMe_3IE_00002.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9406 + }, + { + "path": "EeBKN1mFOeA_00011-Fl610Rsi2Dw_00003.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9407 + }, + { + "path": "OxsvClqjWbo_00010-ihBB7_GYohM_00001.wav", + "question": "How many people are delivering this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9408 + }, + { + "path": "dhy9amxTK4M_00026-seSGUV6a6DI_00002.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9409 + }, + { + "path": "8CDgZ3fZ9mk_00002-lIDDfgKsR0I_00004.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9410 + }, + { + "path": "x3QoGpHxThw_00003-o5cJCp8wF0g_00001.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9411 + }, + { + "path": "kexwwMqJfQw_00003-bLiyGJiyAig_00005.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9412 + }, + { + "path": "58nN00hr508_00005-SSBRk-_tTDw_00001.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9413 + }, + { + "path": "t96KuRu0jz4_00008-rWaBVelzaFk_00017.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9414 + }, + { + "path": "F5aPBZ3PFZg_00010-RLKKsYiCMvc_00029.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9415 + }, + { + "path": "F4fjHj9uppQ_00008-W7680q25xY4_00018.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9416 + }, + { + "path": "0XDrW1o3cgc_00001-p6wKxdg0p9w_00014.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9417 + }, + { + "path": "8iNMKzjk6Lc_00023-aTPLiT8CvPQ_00003.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9418 + }, + { + "path": "gjFhr6ECQzk_00001-KZIY_Rpi0ug_00012.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9419 + }, + { + "path": "vaK4t1-WD4M_00011-JXk3o_ppHqo_00013.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9420 + }, + { + "path": "jUSC4i_eGHs_00004-FPq99qjqaSc_00008.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9421 + }, + { + "path": "yW4-PafmcZw_00003-FQ0rnbBMCqQ_00006.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9422 + }, + { + "path": "3vWez3baO60_00012-HrA2BLcLApA_00001.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9423 + }, + { + "path": "DxZaPGUIX50_00044-GBCgYlOv1ug_00018.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9424 + }, + { + "path": "v5fc53M4QE8_00008-2CFHHsxtG4U_00006.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9425 + }, + { + "path": "hGL2tRHivdw_00001-4y5JogtS8Ek_00035.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9426 + }, + { + "path": "6Hm8SeN2Wp8_00005-O06Emd2vKhw_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9427 + }, + { + "path": "n7D4oSUvDv0_00007-qFNujk5-6nQ_00002.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9428 + }, + { + "path": "0PGoO6SNLjg_00001-hcOFWeQH1TE_00001.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9429 + }, + { + "path": "wQE4ssF-vhQ_00008-5cu92kGSGCY_00010.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9430 + }, + { + "path": "E7qpHfvZnf0_00001-ST2JZXcyYkE_00004.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9431 + }, + { + "path": "iRfu_rotO_c_00010-78qYS9Da15I_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9432 + }, + { + "path": "xZt_yS636rY_00005-ZCY9rFPmLkE_00004.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9433 + }, + { + "path": "94D7hcvA3pQ_00002-7_h0_dHd7_w_00030.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9434 + }, + { + "path": "tWreIBz9lA4_00051-j2r_6XsCMtQ_00002.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9435 + }, + { + "path": "IJJJjQY14u8_00002-5ssVY9a5X-M_00003.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9436 + }, + { + "path": "VDJi9gg6RhI_00001-6qFnVechX9o_00018.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9437 + }, + { + "path": "3uNEf8wV728_00066-o0p1v4IdsAA_00003.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9438 + }, + { + "path": "ccJCCXUVbjc_00004-tFlL3XPi-O4_00009.wav", + "question": "How many people are delivering this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9439 + }, + { + "path": "fJfNJ8T2ECw_00009-GtAHVLPqHLc_00008.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9440 + }, + { + "path": "7P4E-933KyY_00002-dP82yj93J1I_00010.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9441 + }, + { + "path": "NEJ2ekbI9H8_00004-zD7I-4oZ2cU_00018.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9442 + }, + { + "path": "mvs5K53M1Ss_00001-ZOFJhtebvuU_00010.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9443 + }, + { + "path": "GCRKRbZwk6E_00001-CkBmwOzFP0I_00003.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9444 + }, + { + "path": "zqDIe7b19fM_00002-KAR9QkES6wQ_00002.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9445 + }, + { + "path": "RyzLr52ZpdA_00010-ABmQ9_PdR8U_00005.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9446 + }, + { + "path": "cnzoLtP8scQ_00001-m4lBe6sMrxc_00001.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9447 + }, + { + "path": "3OorK7HJaSk_00020-UiQDEiQA_Go_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9448 + }, + { + "path": "eWvntayRfe8_00034-vDOMy98R9hk_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9449 + }, + { + "path": "TkVbedVcwos_00006-v8M4yVx0m_0_00013.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9450 + }, + { + "path": "oNUM8zMr044_00001-NAVk5wmkiac_00002.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9451 + }, + { + "path": "vm7TQNQec34_00005-Ls-5ovSl-cM_00008.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9452 + }, + { + "path": "xtvtXOiqfJ4_00004-0-Lu3WGSfaU_00010.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9453 + }, + { + "path": "TL47aWipItI_00006-lWOn2zc-3dU_00016.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9454 + }, + { + "path": "vfjbbp1sQ9o_00005-vfjbbp1sQ9o_00006.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9455 + }, + { + "path": "UHd9x0J8N7Q_00004-GDskkSMYPMo_00001.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9456 + }, + { + "path": "aVmHBUeThTQ_00001-wIro05OMe2w_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9457 + }, + { + "path": "3lNl1bJt7kw_00009-Z80CH9zm_ek_00002.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9458 + }, + { + "path": "HAOpZNjU_H4_00009-VTt_jQv2I8s_00005.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9459 + }, + { + "path": "JNjPvyEoGXU_00010-gZyyaH3ITq8_00013.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9460 + }, + { + "path": "nC5Sk99wujE_00001-YBE9X_mpy60_00002.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9461 + }, + { + "path": "B5-g7AWjMCA_00001-geHj6NzG99E_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9462 + }, + { + "path": "YSHJrM6PFFE_00005-k_BzHVq5qg4_00004.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9463 + }, + { + "path": "C54L2kqxQoA_00003-qYMOwzv4Lpg_00002.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9464 + }, + { + "path": "yD6IahyNlgk_00010-rstp60Rzgy0_00015.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9465 + }, + { + "path": "HJnzTUM9G8E_00001-RNKUA8EGl20_00010.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9466 + }, + { + "path": "GoBIaYQRmh8_00001-Qk2iYimTmIs_00006.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9467 + }, + { + "path": "q2BoSBDt3qo_00038-q2BoSBDt3qo_00010.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9468 + }, + { + "path": "irFX_c9HQlY_00001-09H3uJc9fI8_00001.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9469 + }, + { + "path": "1eQRShYbQXI_00006-KqlqyYGuOlY_00010.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9470 + }, + { + "path": "1yc4hUIgyKA_00009-Ngmd5dadQkc_00012.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9471 + }, + { + "path": "ed_fH0Jddr0_00003-6clajJmjjPE_00001.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9472 + }, + { + "path": "jQnql_1gbXA_00004-haaRqtnZwMI_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9473 + }, + { + "path": "X-wtgXmhT1o_00004-gQgj-fB7jzw_00010.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9474 + }, + { + "path": "b6INQiSm7pQ_00002-230QPHWW7fM_00017.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9475 + }, + { + "path": "D-MxmWfoJ7k_00005-C_Yw2UQyKR0_00001.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9476 + }, + { + "path": "CdYDTiK33l4_00003-_hDTlD9ApP8_00004.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9477 + }, + { + "path": "Lac6z5aKqTI_00002-PP0cQLwYz5M_00003.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9478 + }, + { + "path": "mmJdPAm7YKo_00009-52QBJw9U_3c_00001.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9479 + }, + { + "path": "6W9VRE-gj3s_00003-QiP4_B02Nr8_00035.wav", + "question": "How many people are delivering this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9480 + }, + { + "path": "JmnX5IIpNPs_00004-R1UIBckH-no_00009.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9481 + }, + { + "path": "DxZaPGUIX50_00160-8BlOgFYSAG8_00007.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9482 + }, + { + "path": "AXjkhN1FXfI_00002-HNcyYJlxWJg_00004.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9483 + }, + { + "path": "zqAvB_cGfko_00020-ljWBcAneMC4_00002.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9484 + }, + { + "path": "ESIjgUlVEP0_00005-Pz9np_gxN6E_00005.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9485 + }, + { + "path": "xcXoT1frQr0_00017-bC6tUNIR9V0_00077.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9486 + }, + { + "path": "2GCCSDEPRNs_00011-g5Z0ygD4fa8_00014.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9487 + }, + { + "path": "_xX4p2wGftQ_00002-qgQssXEXgyA_00007.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9488 + }, + { + "path": "q60ulj4fbyU_00006-LgMcDT7PKuo_00001.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9489 + }, + { + "path": "wMPjKZ4vQec_00012-dNMZWJ6k5J8_00002.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9490 + }, + { + "path": "ZkGb4cY9iRA_00004-ozkcxdYxoRI_00008.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9491 + }, + { + "path": "7yx9A0yzLYk_00012-7yx9A0yzLYk_00007.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9492 + }, + { + "path": "RAy_AWTdZtg_00007-6KjsxA94YI0_00010.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9493 + }, + { + "path": "h1pY451KF7E_00001-AmGedZsxPW4_00001.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9494 + }, + { + "path": "L0sMAaA2C14_00002-5I6U6j3tomQ_00010.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9495 + }, + { + "path": "ni-UxcEZqB8_00004-9fnYJcEW0J8_00013.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9496 + }, + { + "path": "DlfzQltb7cM_00001-UN6BeUs8td4_00020.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9497 + }, + { + "path": "ELuoYAC01rI_00005-3uba8f38NiI_00004.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9498 + }, + { + "path": "IeRkCV_1jmk_00001-Wh3hZmPPWLc_00003.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9499 + }, + { + "path": "shIRpv154g8_00006-KPYCxRWp2sM_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9500 + }, + { + "path": "K38Ya_Sag_g_00001-v4_I-SPHI_Q_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9501 + }, + { + "path": "hUSIudsLXLE_00001-0x6UxKMetG8_00010.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9502 + }, + { + "path": "VtaVbKqxbsI_00011-APrARrLD03w_00015.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9503 + }, + { + "path": "z0B4KEXVaqo_00002-YcJq-UXGRNw_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9504 + }, + { + "path": "VRE-IoEOC0g_00005-zXLJTCRdMFE_00001.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9505 + }, + { + "path": "lT7RLnELEds_00003-ttM_RuPsbS8_00005.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9506 + }, + { + "path": "w2aGGP2-xFo_00006-KroZTXSByAs_00003.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9507 + }, + { + "path": "wgoAh9ji-Ps_00009-Df5SU_3RPtE_00005.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9508 + }, + { + "path": "uqJOS2DkgmM_00002-AnELf0dngpA_00004.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9509 + }, + { + "path": "dKv7e96RdhU_00006-qqXwNd3dcgM_00001.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9510 + }, + { + "path": "eTg_ONDWG10_00001-VNowySPXrMc_00004.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9511 + }, + { + "path": "K4OJq-PpZoI_00018-hO1o58zKJfE_00002.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9512 + }, + { + "path": "kxMo_3J5X9E_00001-veRBeA6noWM_00002.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9513 + }, + { + "path": "DRDn6NFK9Gs_00022-40umohtuhMc_00002.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9514 + }, + { + "path": "B5EVNvjGTI8_00005-P0OKdQWm088_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9515 + }, + { + "path": "yvFQpExofZo_00013-Qt7XmcD6akY_00015.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9516 + }, + { + "path": "n4g3yDkAj00_00002-ryI2SFdMlkg_00002.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9517 + }, + { + "path": "xjDWaXvBtVg_00007-u3Lza0XsXnE_00004.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9518 + }, + { + "path": "Z_b52yo0iQU_00007-qM5FO3hHEbk_00029.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9519 + }, + { + "path": "K3VF9KATPqc_00051-P8bFDbigcYY_00008.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9520 + }, + { + "path": "RFF0JbkRu38_00006-8GImi3iNPJ0_00001.wav", + "question": "The speech captures how many different voices?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9521 + }, + { + "path": "9loouKtho-Y_00002-tc3mgi3td7k_00005.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9522 + }, + { + "path": "AUQRp8D66hE_00004-YcJq-UXGRNw_00003.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9523 + }, + { + "path": "NuokWhpheSk_00008-QI97SKO4dYI_00001.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9524 + }, + { + "path": "E_XuNSZrfY8_00004-yVPbCIFALAU_00002.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9525 + }, + { + "path": "vZjpmzzZ_ng_00004-zJn0HwsZkCU_00003.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9526 + }, + { + "path": "U0bIojE3spc_00005-STRWGiLegRQ_00009.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9527 + }, + { + "path": "0FHidop4Cyc_00010-wy1oXdoq1hs_00004.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9528 + }, + { + "path": "lc1yDLHEKT8_00004-fvc2VO3JNSs_00003.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9529 + }, + { + "path": "pIKl9HcUymk_00029-KIXsd1t7PYA_00003.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9530 + }, + { + "path": "sLZWg2Xs_5w_00002-uBfryYcxhn4_00004.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9531 + }, + { + "path": "i_4R28RjmJo_00003-IrEtrjDedX4_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9532 + }, + { + "path": "5bE7zSyjNSs_00085-poSJrdIP3xY_00005.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9533 + }, + { + "path": "OS1I7a4LNww_00011-kWJBSuuqnwE_00005.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9534 + }, + { + "path": "McLz77opBog_00011-4IrzgonbcPs_00006.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9535 + }, + { + "path": "d3o1nZ-UsAs_00028-Diyrk9giFvU_00008.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9536 + }, + { + "path": "gCYRi_3rqm0_00001-UE1d3bzmJz4_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9537 + }, + { + "path": "XR0sKT4Qvgk_00002-zEoKy-VRE5U_00001.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9538 + }, + { + "path": "ztrh92ku3k8_00012-pBnUi0uiM7Q_00003.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9539 + }, + { + "path": "YsYAmfH46Ys_00006-OOEdY4IuQ5Q_00007.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9540 + }, + { + "path": "8pDGaKmyWhU_00014-ZWePlXe0sG8_00018.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9541 + }, + { + "path": "M0acERfJwgw_00009-DLQY19QumUc_00050.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9542 + }, + { + "path": "RMDXpJ5e_kE_00150-FP_vdf3MkNY_00001.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9543 + }, + { + "path": "2iZQFpRit3E_00005-nQZLZ8uoddE_00007.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9544 + }, + { + "path": "7nRRPPR6XlY_00002-_LCjFvDXK-k_00003.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9545 + }, + { + "path": "1O5YJIGl1X4_00013-Yf93Q8L4xhs_00011.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9546 + }, + { + "path": "fyWvADteroA_00006-DOhJFBO3cd4_00005.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9547 + }, + { + "path": "aBH7pmHztTQ_00004-9GcxwgvzJxk_00003.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9548 + }, + { + "path": "fe7Ln62O9GU_00027-5fUxYQUBtLA_00014.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9549 + }, + { + "path": "ssFfrN2NQxA_00053-KqRnDF7Rqks_00007.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9550 + }, + { + "path": "vvLu_7TpP8c_00028-J8hy2i39_q0_00001.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9551 + }, + { + "path": "abZUA8TQroY_00007-NAsc8wjQouI_00007.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9552 + }, + { + "path": "RDFdX3VxjUQ_00023-0sabETmKH5w_00002.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9553 + }, + { + "path": "y5JWex5vBkQ_00002-35-qPdnh8BY_00002.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9554 + }, + { + "path": "j3UYusVHkp4_00001-DMBems4y_p0_00004.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9555 + }, + { + "path": "O77p-6-AqP4_00046-wMFu6MBG_JI_00007.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9556 + }, + { + "path": "BLMrBnY_sNI_00004-rv_Xkz9zepI_00039.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9557 + }, + { + "path": "UmCt1yhIaZY_00013-8kYpZukmgKY_00002.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9558 + }, + { + "path": "DSwqIT0_cgI_00007-YSaI6OUVli8_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9559 + }, + { + "path": "BFqLRtj_Nh4_00006-Ns7ocpRhDD8_00001.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9560 + }, + { + "path": "Cij1T-c3ci8_00044-ZeX74Xx9ZYI_00001.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9561 + }, + { + "path": "IQa3kkKNtgA_00002-89FrzqI2x0A_00006.wav", + "question": "The speech features how many speakers?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9562 + }, + { + "path": "Q-zImmcBLoQ_00019-a_GeH7Tnzag_00099.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9563 + }, + { + "path": "O1TR8kzqKcA_00004-MYfw_v7Cqcg_00029.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9564 + }, + { + "path": "fBamkF0lp_M_00002-YYl7evawZnw_00008.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9565 + }, + { + "path": "tvfx-OlFisE_00010-q7xxJIiRNoM_00003.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9566 + }, + { + "path": "GE-LQwmgX0c_00012-IlhmVLEwysA_00015.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9567 + }, + { + "path": "jBpVO9pLeUI_00009-4Bz1ThNOONQ_00003.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9568 + }, + { + "path": "pQMUErWp-CA_00005-2ZBmCoHGZ8k_00005.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9569 + }, + { + "path": "8hMkhwBzaJY_00004-UezhcWG-HRY_00002.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9570 + }, + { + "path": "ADLO9eWGUIY_00024-uNdfTrSvmqY_00002.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9571 + }, + { + "path": "dYTzceCBQxQ_00006-8e-XM55GGsk_00015.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9572 + }, + { + "path": "ve6qA-6_Q5g_00018-UEQzqaQNTMU_00002.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9573 + }, + { + "path": "wn6RACKwA_A_00001-D1lwqtw-hMo_00003.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9574 + }, + { + "path": "Blt_H-Qq1jM_00027-GSxuvuI8fNo_00003.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9575 + }, + { + "path": "q47gPbZG6wo_00004-cRESGIrJQoc_00001.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9576 + }, + { + "path": "XhaEuVtX4cg_00006-Mv8ZKqiiv0k_00004.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9577 + }, + { + "path": "zmVMzOWUzko_00003-oOGfyKHy4O8_00004.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9578 + }, + { + "path": "at-3FO1kIC4_00018-at-3FO1kIC4_00018.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9579 + }, + { + "path": "s_rtHBpzrQc_00008-5nUwVGiD2JY_00010.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9580 + }, + { + "path": "SD8tdAWYKy0_00007-rEhBM_VI-Ms_00008.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9581 + }, + { + "path": "SMeBRz5AKDY_00001-EP5e91r30OI_00013.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9582 + }, + { + "path": "1C2dwye3Jzk_00020-IXZKoUzkvc0_00017.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9583 + }, + { + "path": "5PBimrKDrJQ_00001-355AMVyCyIg_00008.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9584 + }, + { + "path": "FtEJKQ6zBXI_00001-LD4xa_UEcxc_00016.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9585 + }, + { + "path": "ZCY9rFPmLkE_00005-SK9iVw56wXs_00015.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9586 + }, + { + "path": "zuOySOcZM2w_00009-4eStFmZMSkg_00072.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9587 + }, + { + "path": "pU-d_RvIvIQ_00002-ve6qA-6_Q5g_00027.wav", + "question": "The speech captures how many different voices?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9588 + }, + { + "path": "4JZlYI-fR4E_00008-jyTYRG0Hb4A_00018.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9589 + }, + { + "path": "jkyyjMqdck4_00007-oVJhNZO_M9Q_00040.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9590 + }, + { + "path": "Q-zImmcBLoQ_00010-U8FIDkp_kJI_00020.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9591 + }, + { + "path": "xho_C_k6lWk_00013-eipT9BRvqYE_00001.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9592 + }, + { + "path": "nHHR2PDyr3Q_00009-HsBs2i8C1G4_00007.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9593 + }, + { + "path": "WcLzSLJ8RB8_00049-dYK2s3YM_o8_00024.wav", + "question": "How many people are delivering this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9594 + }, + { + "path": "r9-0pljhZqs_00007-eC9ZIOvohds_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9595 + }, + { + "path": "PeuZp2Tz0rc_00001-KrRwmRAce2Q_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9596 + }, + { + "path": "jES63z_THgQ_00001-NP3WmxUdvnI_00005.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9597 + }, + { + "path": "9v7nAzw5lXI_00017-RjlNX9tYC6M_00042.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9598 + }, + { + "path": "Rx291xS79aQ_00001-TEa-X67ciGY_00013.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9599 + }, + { + "path": "Q2E0FalJFLU_00023-UOxg8l4ykPQ_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9600 + }, + { + "path": "LyUxrJtnsuY_00005-XrRyHHWa4ew_00002.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9601 + }, + { + "path": "8J1CnIq9uIo_00042-ymRDXj_P9T4_00004.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9602 + }, + { + "path": "th0v11Btk1I_00027-eCA5o-Nnfuo_00012.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9603 + }, + { + "path": "UlrmRKDIdTc_00011-UHd9x0J8N7Q_00011.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9604 + }, + { + "path": "kYJi779BvMU_00002-6POKNa7L3uM_00007.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9605 + }, + { + "path": "RCMCIv60_Rk_00002-TmMvlOgvPyQ_00004.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9606 + }, + { + "path": "0jA-z5OgFgk_00024-2KiIeExNAsM_00038.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9607 + }, + { + "path": "MbKOGipoV7Y_00004-6o-LScd5k38_00004.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9608 + }, + { + "path": "2Er-cV85fc4_00027-l3f4NFspWuc_00004.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9609 + }, + { + "path": "qgQssXEXgyA_00001-k_Pt6x6zl70_00001.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9610 + }, + { + "path": "qQCi5-78tIs_00003-SStbmbmTU30_00013.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9611 + }, + { + "path": "0_fWUKLUAb8_00001-45HMy-p0GP8_00003.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9612 + }, + { + "path": "pwjdQHAaoyc_00001-wEO3blmXr6E_00003.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9613 + }, + { + "path": "Jbq5MoUBaXU_00001-hcOFWeQH1TE_00003.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9614 + }, + { + "path": "8Jim3lamzDQ_00004-LQQ1COemsOM_00003.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9615 + }, + { + "path": "DWGw-h0-xyA_00003-i6SqsB3OsBE_00011.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9616 + }, + { + "path": "TAIgNJGh_lE_00012-E_T90bDDDxU_00001.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9617 + }, + { + "path": "T5VqQSfzmHo_00005-CHwD2nAP4OU_00002.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9618 + }, + { + "path": "6D6GMQoOImI_00002-ONy-ZeEkMsk_00004.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9619 + }, + { + "path": "9lXaK5ExJtU_00003-m0pirbsgB-c_00092.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9620 + }, + { + "path": "bwYAk5HVPCk_00001-WaqVW9NUjzA_00003.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9621 + }, + { + "path": "90mWMOAv0wk_00001-WG0WzFM5IMs_00005.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9622 + }, + { + "path": "cRYaG6y9TmU_00065-OwiGYNpf6Pc_00002.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9623 + }, + { + "path": "fI-1LgrsNNE_00022-DOJkuy_S-a8_00003.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9624 + }, + { + "path": "NcIFNlN2yXI_00006-OFb2NfMOekA_00004.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9625 + }, + { + "path": "ZoHKX2DKvIE_00004-ZUfzXz23ndo_00005.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9626 + }, + { + "path": "lBHpsPXdef4_00003-zTgjHcTZ4ic_00002.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9627 + }, + { + "path": "sIo3tYrd_5w_00004-ObvIB1teaa0_00018.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9628 + }, + { + "path": "bTpWDPldduE_00001-DxZaPGUIX50_00030.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9629 + }, + { + "path": "LNtxkQZHgjg_00014-LNtxkQZHgjg_00004.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9630 + }, + { + "path": "pJ4Jnsz1SsA_00014-SK9iVw56wXs_00003.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9631 + }, + { + "path": "p1SQEcYYEI4_00010-tUoSAxOAzSw_00004.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9632 + }, + { + "path": "Jx8GYQSmVJM_00002-qlMFZfWY-PE_00003.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9633 + }, + { + "path": "EgmXPVgnxMA_00005-G0eh4mutyzA_00006.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9634 + }, + { + "path": "q7xxJIiRNoM_00057-1uUxa00zKXE_00011.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9635 + }, + { + "path": "E3J23SXNUas_00004-qB4ltWRf34Y_00006.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9636 + }, + { + "path": "nDO1L-HZGI4_00011-l4HDjqZxAHQ_00009.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9637 + }, + { + "path": "t9pbDI-qFIw_00001-U8FIDkp_kJI_00019.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9638 + }, + { + "path": "CrUpSXxGEXs_00005-P1uwrb8YZaw_00003.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9639 + }, + { + "path": "kSOvDX4n_IE_00002-knYyzOkKyM4_00011.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9640 + }, + { + "path": "wSvoGvXQZug_00020-tRV9eCFKj_k_00002.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9641 + }, + { + "path": "RjlNX9tYC6M_00048-UX3smaRv7-Q_00004.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9642 + }, + { + "path": "WxmrMgdkqOw_00003-JjQwYKMSq-Y_00001.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9643 + }, + { + "path": "G0tMaDPU1h0_00002-NMdgNG1MciI_00003.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9644 + }, + { + "path": "VFyPhJmxazI_00046-CIDyNTLY2Y0_00003.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9645 + }, + { + "path": "ODo_jw4ezjM_00001-l7FyyYZ0U5o_00001.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9646 + }, + { + "path": "8vAjqT9yWRQ_00001-08755wOq7aM_00021.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9647 + }, + { + "path": "SY_FpBvyoKE_00013-maPGe4wglNg_00020.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9648 + }, + { + "path": "K7ySNaja0ZM_00009-M0cZnVSw6s8_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9649 + }, + { + "path": "k6vtSSSWTgI_00010-ZoNuup6qheA_00003.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9650 + }, + { + "path": "dbLnn3HYogI_00032-RW4U4bB0AMY_00003.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9651 + }, + { + "path": "dzxjJoRMpVw_00020-Ts9BtNw-Y74_00004.wav", + "question": "How many people are delivering this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9652 + }, + { + "path": "ngODTKKNnZU_00007-lPCbzQKGBLI_00008.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9653 + }, + { + "path": "qq5QgXy1-80_00005-XJq32ejsjk0_00007.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9654 + }, + { + "path": "PpIVqWf_JwQ_00005-RsvVf8bCFkI_00002.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9655 + }, + { + "path": "dL24Btmisto_00001-l6ZF3b7LrYo_00003.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9656 + }, + { + "path": "6iBO5AjpvOQ_00006-HaA9mTYzZgs_00009.wav", + "question": "How many people are delivering this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9657 + }, + { + "path": "lr_kiG1MLhI_00001-TxlwLr4b4E0_00002.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9658 + }, + { + "path": "6J29pHzTxN0_00003-34k7u_wBZi0_00002.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9659 + }, + { + "path": "kAh-Yd0Rmdw_00010-IHGVJEvRo-c_00012.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9660 + }, + { + "path": "UEsNmbiwIb4_00002-J_TurwsR55o_00007.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9661 + }, + { + "path": "tiq90I2YObs_00013-sh6J3wEmceA_00005.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9662 + }, + { + "path": "acuXfGycBHw_00003-1BN1Twr0pDM_00004.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9663 + }, + { + "path": "zuOySOcZM2w_00007-u7qCFBP1nH4_00005.wav", + "question": "How many people are delivering this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9664 + }, + { + "path": "RAbQ3ujCdQQ_00010-WcLzSLJ8RB8_00067.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9665 + }, + { + "path": "z7Vn-WzgYAM_00001-8irGzD7h_PE_00019.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9666 + }, + { + "path": "vy8sQ82o0fM_00007-sXapU-vWaNE_00038.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9667 + }, + { + "path": "wIro05OMe2w_00002-WB20VHYBO6o_00005.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9668 + }, + { + "path": "qFNujk5-6nQ_00002-i6YSVo-OP0M_00025.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9669 + }, + { + "path": "dV8Q5-tEB6A_00003-fF5qt1bEgPo_00012.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9670 + }, + { + "path": "gg-JjtpYqBQ_00002-gg-JjtpYqBQ_00001.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9671 + }, + { + "path": "O6qLnZfCsdE_00006-kYkGm1VPenU_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9672 + }, + { + "path": "LFV2pleiv5k_00002-D51auzYLESw_00007.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9673 + }, + { + "path": "Yocizg57t48_00009-OSUGWP3iYns_00009.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9674 + }, + { + "path": "DbN5tyZ5IvE_00006-VRqtsKJVthk_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9675 + }, + { + "path": "qupY4hAS6h4_00007-I9gCyrZWFn0_00015.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9676 + }, + { + "path": "q3xmwy0T9Ro_00004-LZcuJkDTO2s_00002.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9677 + }, + { + "path": "gb0n8ZsYw58_00002-8l5ZnDf-FUA_00020.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9678 + }, + { + "path": "csbEEsKuRO8_00002-SIQ2HQIcMUU_00009.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9679 + }, + { + "path": "T42h1PKwQBg_00006-z7VKNevTgl4_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9680 + }, + { + "path": "JyzSALz3TYk_00014-FeQ9b5bMf9c_00007.wav", + "question": "The speech captures how many different voices?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9681 + }, + { + "path": "p6xwhCREA94_00003-OK82eJJwCGM_00001.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9682 + }, + { + "path": "XPnrUmxql7Y_00026-MV4B_V28-j8_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9683 + }, + { + "path": "QtaX-y85YnM_00004-VMpl45SFqC8_00004.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9684 + }, + { + "path": "skKjjUaCUq0_00005-zbs08XbNC3c_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9685 + }, + { + "path": "noB1BoL5vT8_00002-vfUzWn_R7jE_00007.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9686 + }, + { + "path": "0K2_aMAeo-s_00019-5-6lI5JQtb8_00002.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9687 + }, + { + "path": "pxUF7wuXtMQ_00001-hjvQiiG71rM_00021.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9688 + }, + { + "path": "wfBk-ezUiVs_00006-746K5NHwYmA_00002.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9689 + }, + { + "path": "CQDQQyxHmzI_00005-vVKV04EwOEc_00002.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9690 + }, + { + "path": "CN6JpPYbHCQ_00006-qFE4ZVeSRU0_00012.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9691 + }, + { + "path": "mpf2NYnPTNE_00004-iNhHCcePFPI_00009.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9692 + }, + { + "path": "zii-zlYDpGc_00007-cIi74VD0LF0_00002.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9693 + }, + { + "path": "ZEpT2orkZTc_00001-4tlCKgb3LvU_00003.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9694 + }, + { + "path": "vSgHzBY61_o_00008-beMhG6PENYI_00004.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9695 + }, + { + "path": "5ad_KUPOvLc_00013-Qv7RVdd9LtI_00004.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9696 + }, + { + "path": "hYJwTCeuTfU_00003-AR0804ul8mg_00001.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9697 + }, + { + "path": "rmwG4aQfKnw_00006-idYKYA_HEco_00002.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9698 + }, + { + "path": "uw4TsuWHBjE_00026-5V38qDlvxPc_00007.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9699 + }, + { + "path": "NoAA12VcG0I_00002-sh6J3wEmceA_00013.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9700 + }, + { + "path": "SxoxgMD7XTY_00001-cUIYzhAWtEo_00012.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9701 + }, + { + "path": "8QNLnekKaPU_00017-8QNLnekKaPU_00011.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9702 + }, + { + "path": "DxZaPGUIX50_00116-1R3XTTBorKk_00005.wav", + "question": "How many people are delivering this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9703 + }, + { + "path": "3hKKkGhEDoU_00006-qBaOrJpbMTs_00004.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9704 + }, + { + "path": "_mUc_Xl3gdM_00034-qKM9EwcfLsw_00009.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9705 + }, + { + "path": "k7KHJHQDxDA_00005-r5aBZ6SIzR8_00006.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9706 + }, + { + "path": "ABmQ9_PdR8U_00005-BBu8a23kbV0_00003.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9707 + }, + { + "path": "HYKhm9ulxSU_00002-AUQRp8D66hE_00011.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9708 + }, + { + "path": "Ltttsep1HC8_00008-uNO8DynLFZo_00001.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9709 + }, + { + "path": "SkXo5pAoP3A_00018-xtua_Uq8kUU_00018.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9710 + }, + { + "path": "JBl5rJuQKRo_00003-JBl5rJuQKRo_00015.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9711 + }, + { + "path": "ZslJpHH4bQ0_00002-gdomtGoYaa4_00002.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9712 + }, + { + "path": "4HMR-DEO8WE_00006-RHgFtU9eAlA_00003.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9713 + }, + { + "path": "XsiRQ8XiWOk_00002-cNKDSeXivmQ_00029.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9714 + }, + { + "path": "8J1CnIq9uIo_00003-HvBwJrc_-ns_00002.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9715 + }, + { + "path": "h5Regts09Fs_00002-8-VJB3Je_d0_00005.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9716 + }, + { + "path": "I8VJBexmRIQ_00004-MTHJaGEJ_L8_00017.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9717 + }, + { + "path": "jI2mcN6F6SA_00009-qFUUDSjCy5M_00003.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9718 + }, + { + "path": "cAk31IW1GZI_00004-FfKvI_HhRmM_00002.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9719 + }, + { + "path": "EGVcQhK3b5o_00054-BhLxC_Ypzew_00023.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9720 + }, + { + "path": "6epiNL5b4wo_00001-dZDFZaR4CeU_00016.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9721 + }, + { + "path": "lSI69L4DidQ_00003-wKpYLlZVGrI_00003.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9722 + }, + { + "path": "Wvsd1vi7rJw_00001-Wvsd1vi7rJw_00001.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9723 + }, + { + "path": "YloBWJxXzMI_00013-ChJ5k97QI5g_00005.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9724 + }, + { + "path": "YkWjfSi9Jrg_00002-Bvr5NarHLRc_00014.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9725 + }, + { + "path": "uKJKBqC-Roc_00010-v19q5ycEXH0_00002.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9726 + }, + { + "path": "YOjB3bkUrUA_00002-TrSRgOom4qE_00005.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9727 + }, + { + "path": "A_blsQbRRC8_00001-HELG9IIqhBE_00005.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9728 + }, + { + "path": "CRP3XtjNdeE_00009-K0il0nc1y3U_00009.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9729 + }, + { + "path": "VFyPhJmxazI_00046-3KkpGq8U4KE_00002.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9730 + }, + { + "path": "slsMCVL9qlg_00002-qxG38OMQ1W4_00007.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9731 + }, + { + "path": "dg_LUd235gk_00021-hE-gAqJwkTI_00001.wav", + "question": "The speech captures how many different voices?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9732 + }, + { + "path": "dQhROKhai8Y_00009-WBjWoYv8WcI_00010.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9733 + }, + { + "path": "j3RAP5Rzss8_00003-LD1b8PNHaAg_00002.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9734 + }, + { + "path": "U3xR3MZjEVg_00005-PIorPTaj_Rg_00005.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9735 + }, + { + "path": "UZZlFl9P22Y_00002-ewtrj4bTGh4_00003.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9736 + }, + { + "path": "t2b-CAsCbkc_00017-N4q9dIbw68Y_00006.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9737 + }, + { + "path": "bnjgG7dzIfY_00007-d9B3S3Lm208_00001.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9738 + }, + { + "path": "G2QZRjUB_VM_00087-aGTbe7lffgg_00001.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9739 + }, + { + "path": "FWGmZq_wkgY_00003-n6j-UuSijgc_00008.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9740 + }, + { + "path": "jCpG_wDhjcM_00004-v60aXisPGIc_00009.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9741 + }, + { + "path": "661UcMkDQJ8_00006-XItoro6xkB8_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9742 + }, + { + "path": "U58dh2nGIsg_00003-NKJp3rNrMnY_00032.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9743 + }, + { + "path": "xOlm69WCj4M_00006-ZCTpP0ECz6k_00010.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9744 + }, + { + "path": "ToOr53u4u_A_00004-TDLxOdF2KUQ_00002.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9745 + }, + { + "path": "K2QHfrmf56k_00055-HYiqXC0B3iY_00023.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9746 + }, + { + "path": "LuYBLVA7K58_00003-hXmtIVGVBEo_00011.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9747 + }, + { + "path": "nFPWqjTyqTo_00004-Ok_1huCJP0o_00018.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9748 + }, + { + "path": "DL0lXV30xOI_00002-g9X-GTkYcaQ_00005.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9749 + }, + { + "path": "dINKWWFmMyo_00054-HhGAogrgF5E_00027.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9750 + }, + { + "path": "3tCsEFnEO-M_00009-I_jUkSDfAz4_00002.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9751 + }, + { + "path": "UIbPolhxLDI_00004-znxUWA2QAGs_00001.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9752 + }, + { + "path": "b3JpIFdWUVk_00002-CfEVLzd5j7Q_00001.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9753 + }, + { + "path": "BfgM3ER0Etg_00003-VB7sZ6phd1w_00005.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9754 + }, + { + "path": "is5z_0Ic5dE_00001-YJ4k27W5k9I_00001.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9755 + }, + { + "path": "IYAzkRRCHkY_00004-b6Iy3-dDS9U_00002.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9756 + }, + { + "path": "ZQ1897M5GRM_00005-4svG40oHL60_00002.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9757 + }, + { + "path": "3kvWhMqqhhA_00016-qYwb7vCJqUc_00011.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9758 + }, + { + "path": "Q0YU_S7jFhA_00095-Kf2sYUn9sj8_00002.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9759 + }, + { + "path": "QGlexcvxGFo_00019-3mgBwCxTRDY_00008.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9760 + }, + { + "path": "_IoMoOPdGmY_00010-P_JnhYrNSYY_00003.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9761 + }, + { + "path": "KvlQtXMwrTI_00001-zuwGkiAtslI_00003.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9762 + }, + { + "path": "Zdmm9Mrr8Ts_00004-2uBsh9dTB8g_00029.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9763 + }, + { + "path": "2TIs7b26obM_00006-fyOSQW-ETeE_00002.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9764 + }, + { + "path": "wJ_YINYHBwY_00012-4y5JogtS8Ek_00018.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9765 + }, + { + "path": "yp307tVx1jU_00014-bXpavyiCu10_00015.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9766 + }, + { + "path": "dg_LUd235gk_00039-aWAguBjO03M_00006.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9767 + }, + { + "path": "4o3GZTwaPvo_00009-u_LvGyrI8K8_00011.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9768 + }, + { + "path": "vCISva6imEc_00001-SWDuFCqLAqQ_00003.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9769 + }, + { + "path": "RAbQ3ujCdQQ_00013-HQkU71i8mlM_00007.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9770 + }, + { + "path": "H5I-FSvSH-g_00025-t4ivCDhpcvw_00079.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9771 + }, + { + "path": "EhRbjkkUygc_00001-hy_wFnrlIoA_00002.wav", + "question": "The speech captures how many different voices?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9772 + }, + { + "path": "7kHcVUZKQdw_00001-7s0XxchGaIU_00017.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9773 + }, + { + "path": "jo-hm_Y9h-0_00001-qqMQ6-FmQbA_00002.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9774 + }, + { + "path": "z4a-lKZ3-f4_00012-QGlexcvxGFo_00011.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9775 + }, + { + "path": "iDxM7uzjVrI_00012-cFv9Ofrmc9k_00001.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9776 + }, + { + "path": "2KiIeExNAsM_00050-aTPLiT8CvPQ_00017.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9777 + }, + { + "path": "tmkV1yS7bDg_00004-zEcPnIzGY_k_00003.wav", + "question": "Enumerate the speakers in this spoken segment.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9778 + }, + { + "path": "DGGTJx470Ag_00003-rxz0tUiutbw_00002.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9779 + }, + { + "path": "iZ6XrudZsSw_00003-LSMOrOJOB2I_00018.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9780 + }, + { + "path": "N0Rznop7WME_00015-WEwMmMkGGlE_00001.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9781 + }, + { + "path": "j4avyGkk2T8_00002-7JFUqzlqhoE_00012.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9782 + }, + { + "path": "0b0i1OMe-3A_00003-0b0i1OMe-3A_00001.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9783 + }, + { + "path": "Z-bPrDRGq2w_00006-6toNHGVS3mk_00005.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9784 + }, + { + "path": "sGqnn4WpQCI_00001-Xf9g2-pCmn8_00006.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9785 + }, + { + "path": "LJ3wryVdDPg_00003-Mnh3W0hr1nc_00003.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9786 + }, + { + "path": "6f7xJ9vFSNo_00019-gLqQYVwlQF8_00003.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9787 + }, + { + "path": "ZML7N9XSfIw_00008-lq2Z0RoiKqc_00008.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9788 + }, + { + "path": "V6Ea4g46FBA_00022-XR0sKT4Qvgk_00001.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9789 + }, + { + "path": "ZAwqKT7EHBg_00003-IvL6Ab33qRQ_00002.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9790 + }, + { + "path": "ut5VuTtFz9o_00001-gnZEyw92hLU_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9791 + }, + { + "path": "E98_GfeZPAs_00001-ZL7wvRSGYXI_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9792 + }, + { + "path": "137siDbXp0I_00007-DxZaPGUIX50_00141.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9793 + }, + { + "path": "JQtDfEz08aU_00022-TSrr2vfXJKA_00001.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9794 + }, + { + "path": "nqytzgB6BYA_00013-bW6NQ7_tDTs_00030.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9795 + }, + { + "path": "Mp1IXRquJ9s_00008-rZQMeH1FCvM_00013.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9796 + }, + { + "path": "ylvffTxPM0U_00001-90sTPYNz4Oc_00003.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9797 + }, + { + "path": "A_d0g4JE-PA_00003-qOjw-X0fxt0_00002.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9798 + }, + { + "path": "0lQ2j2C05PI_00014-oPgdsm4tVDY_00002.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9799 + }, + { + "path": "6SBGrr4iFp0_00002-Sdos4P-O8jY_00005.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9800 + }, + { + "path": "ObvIB1teaa0_00018-tCFleAYicKY_00007.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9801 + }, + { + "path": "1aXDfiBMEwg_00011-m8rWUWniJwI_00001.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9802 + }, + { + "path": "u1iaCkCVTg0_00001-_lNNLktXC3g_00002.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9803 + }, + { + "path": "_ZDNoW_f2qk_00018-EGmlkOP6WMs_00003.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9804 + }, + { + "path": "f-mjLyC1-4g_00001-YONIr2SdpOc_00005.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9805 + }, + { + "path": "DOsLMpphQ_M_00032-dOnuu_9-nyI_00019.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9806 + }, + { + "path": "G9BNoF7zB1Q_00001-aaJo0S1lnhY_00002.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9807 + }, + { + "path": "zv3AoFFDWGs_00013-1lgoknXVPG8_00002.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9808 + }, + { + "path": "tqdtbBSd4HI_00003-EyCIwLrhPKs_00004.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9809 + }, + { + "path": "TfCzhSPY8vM_00003-TfCzhSPY8vM_00010.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9810 + }, + { + "path": "rxMW3ZT6oZc_00007-ciuTCqJkgT0_00021.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9811 + }, + { + "path": "zDkgVesX7NU_00003-OLYJFTb5J2E_00014.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9812 + }, + { + "path": "shIRpv154g8_00002-eyXrm1x6nDw_00009.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9813 + }, + { + "path": "C_qQmRTmRUs_00006-VjAhMRwQY6M_00045.wav", + "question": "The speech features how many speakers?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9814 + }, + { + "path": "jRUrHuZXZCE_00017-K2QHfrmf56k_00003.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9815 + }, + { + "path": "LjLjQkfhd2o_00006-mt5E9pPyxls_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9816 + }, + { + "path": "t24_N3awD-o_00003-eGSQ9aKevhw_00002.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9817 + }, + { + "path": "cc_U4QV8aco_00011-Zs8VK91yVMI_00002.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9818 + }, + { + "path": "4-HouJ8PNSc_00002-yV8NNCEDlek_00073.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9819 + }, + { + "path": "HYiqXC0B3iY_00007-RN6lCy1-D24_00005.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9820 + }, + { + "path": "Xame3EjV0cc_00010-YkaFqsk2yLA_00003.wav", + "question": "The speech captures how many different voices?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9821 + }, + { + "path": "K22AiWKEOg0_00001-4CiHuisRQqA_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9822 + }, + { + "path": "Z4M-Z0AYee8_00009-3oreZ2gxuHA_00001.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9823 + }, + { + "path": "9ADlYZ1Gs6U_00002-jCjAnwcxTDQ_00003.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9824 + }, + { + "path": "ijlt9oQhjgg_00001-9Bu9dYRD28E_00002.wav", + "question": "Count the speakers present in this speech recording.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9825 + }, + { + "path": "1hEr7qKRKL4_00013-Q_q8-k0WRTM_00008.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9826 + }, + { + "path": "EJ6OgzuEJKk_00008-KcrAsV5LInc_00001.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9827 + }, + { + "path": "EI0BcP32ins_00037-etTS-FV1WEQ_00001.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9828 + }, + { + "path": "Ck7SqFpM0DY_00007-ae7wjpp7MIk_00007.wav", + "question": "How many vocal participants are in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9829 + }, + { + "path": "a3ydBB1pilY_00002-OnxRrIhc-l4_00013.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9830 + }, + { + "path": "y-AcEqVpdGU_00008-y-AcEqVpdGU_00020.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9831 + }, + { + "path": "iER54b69dgs_00007-YoUgykRElWw_00013.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9832 + }, + { + "path": "svkE8myNmU8_00001-svkE8myNmU8_00001.wav", + "question": "What is the number of different speakers in this speech excerpt?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9833 + }, + { + "path": "wSvoGvXQZug_00024-7u-rskIpVxE_00002.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9834 + }, + { + "path": "ZaWQnpToL8s_00007-bGHQOzeND2M_00009.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9835 + }, + { + "path": "f6MLIpHjyn8_00008-f6MLIpHjyn8_00014.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9836 + }, + { + "path": "cydHbANANcU_00009-CCbp-IXRYYs_00002.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9837 + }, + { + "path": "A9NxfMke690_00004-RBGhSDz2NZU_00016.wav", + "question": "The number of speaking roles in this speech totals to what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9838 + }, + { + "path": "4PzEhoOxJAU_00030-4PzEhoOxJAU_00007.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9839 + }, + { + "path": "Zw_aTuy8kMw_00002-XsF8YXSV6jM_00008.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9840 + }, + { + "path": "IRuhsRDUjOI_00005-fYCsraSX67k_00001.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9841 + }, + { + "path": "mCtNLv5FwqI_00001-aWTErCL8-ik_00001.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9842 + }, + { + "path": "M7OPlGV4syA_00007-N4QB_ucMxqE_00008.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9843 + }, + { + "path": "zE1eYxkTwAM_00004-6cRnyUtIbJw_00001.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9844 + }, + { + "path": "TnHpZXrLNBI_00010-HbcvFwndLPc_00001.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9845 + }, + { + "path": "7uaGcp151Sw_00007-OLIIuB3hcQw_00006.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9846 + }, + { + "path": "3f7tbKu-s_Q_00001-uNqxF_kBTKA_00014.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9847 + }, + { + "path": "LhTB2hidBqk_00019-zPwcUL9ZLbg_00001.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9848 + }, + { + "path": "020Ug5rAGB8_00004-J9lhAAa_avk_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9849 + }, + { + "path": "tswc1vx5chY_00010-C60tcxZPHuw_00001.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9850 + }, + { + "path": "EBVYy1H_eTM_00003-n4g3yDkAj00_00002.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9851 + }, + { + "path": "NbVrCKilytE_00003-89FrzqI2x0A_00004.wav", + "question": "How many speakers are sharing their messages in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9852 + }, + { + "path": "smYAOqfc7-A_00002-OVOO0f2ZtO0_00019.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9853 + }, + { + "path": "dG53U3yrdCE_00054-Fq9xHBNoPtA_00010.wav", + "question": "How many people are delivering this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9854 + }, + { + "path": "9CgtN03L2z0_00002-xehIp9SaHEY_00005.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9855 + }, + { + "path": "cHbO-tXjg4Y_00003-LTJagcweB_s_00001.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9856 + }, + { + "path": "hlpuBqzI4I8_00001-lDsxhOR3qSg_00002.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9857 + }, + { + "path": "wqMzyy4j--Y_00001-DcAhTUKXx9I_00009.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9858 + }, + { + "path": "k3DGwIlknw8_00003-3MQHAGL6SZs_00002.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9859 + }, + { + "path": "Q_q8-k0WRTM_00005-K7rycYlSZOA_00002.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9860 + }, + { + "path": "kfLrBefnJHY_00001-gQgj-fB7jzw_00012.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9861 + }, + { + "path": "i4GAik0qGgc_00002-73Lz58Oc_1o_00005.wav", + "question": "How many individuals have speaking roles in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9862 + }, + { + "path": "j5uhoHEoylM_00002-2iaCNe7ajII_00003.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9863 + }, + { + "path": "pvEZULqenEw_00008-r4IZRsEGDFs_00014.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9864 + }, + { + "path": "cfbCeo1GDA8_00004-aaebCc11h2I_00007.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9865 + }, + { + "path": "nGWr5F3iKco_00001-E8--f2BFEWI_00003.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9866 + }, + { + "path": "kMXREOtRBsM_00006-gsdbFy3PXdk_00004.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9867 + }, + { + "path": "Awi1Q0yib1s_00008-VRCidrXwd1s_00021.wav", + "question": "The speech captures how many different voices?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9868 + }, + { + "path": "GIV0WD3Ik3o_00001-qwAIKzh6tyA_00003.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9869 + }, + { + "path": "PHV4oAg30DA_00002-ZCGhB9sLeLE_00005.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9870 + }, + { + "path": "NilzUXpSsvk_00009-jJoMRL4i7ns_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9871 + }, + { + "path": "GE-LQwmgX0c_00014-DzOvsPn68J4_00003.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9872 + }, + { + "path": "98bv1J9c9is_00007-ZVzgZVwyvys_00004.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9873 + }, + { + "path": "7JLeel7ok0k_00004-UhX4PNLBNOI_00004.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9874 + }, + { + "path": "zfX-uppaGbQ_00002-sIH47kQ0VJg_00008.wav", + "question": "The speech features how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9875 + }, + { + "path": "DxZaPGUIX50_00033-jZLH__45hVM_00010.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9876 + }, + { + "path": "MKFqfUNnwoI_00001-YZ2gCyx8c2U_00002.wav", + "question": "Estimate the number of speakers involved in this speech.", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9877 + }, + { + "path": "TpG_8C7dzrA_00001-12nv1g3AGUQ_00002.wav", + "question": "What is the cumulative number of speakers in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9878 + }, + { + "path": "96rk6YKwPl8_00003-ntw3-zOROlo_00006.wav", + "question": "What total number of speakers can be detected in this speech?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9879 + }, + { + "path": "MqKFZOAwUsg_00002-Uw3u4E2RjuA_00009.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9880 + }, + { + "path": "RFw8FuLfh7c_00004-tqMxPFzRrGA_00001.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9881 + }, + { + "path": "ggBQPBS9KE8_00006-bbioyrhtKkM_00006.wav", + "question": "The number of speakers delivering this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9882 + }, + { + "path": "vpZ36lIRfmw_00002-tqdtbBSd4HI_00050.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9883 + }, + { + "path": "7uy3nCGHxJA_00003-Ysq3Z4AMTTc_00001.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9884 + }, + { + "path": "o0p1v4IdsAA_00010-yVWutBwaJME_00021.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9885 + }, + { + "path": "8CJKjlhUbQk_00007-agQ0RfUEo14_00013.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9886 + }, + { + "path": "Y-PJPns3-YU_00002-HYiqXC0B3iY_00030.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9887 + }, + { + "path": "QruGwcKVg54_00002-VSrjCY1Ro-g_00011.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9888 + }, + { + "path": "LZQR6zB3xjw_00006-U8rOal2Ty-s_00010.wav", + "question": "What number of speakers take part in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9889 + }, + { + "path": "_BUSj8NIOmw_00043-TX4AzuABquI_00024.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9890 + }, + { + "path": "R1UIBckH-no_00003-UUpSDo9_Qi4_00013.wav", + "question": "How many people are part of the spoken dialogue in this speech?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9891 + }, + { + "path": "MMTrblYIEgM_00001-RREb7K-SoSQ_00021.wav", + "question": "The number of distinct speakers in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9892 + }, + { + "path": "SUEmbQVo6Yc_00003-FkyvXlj-hec_00006.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9893 + }, + { + "path": "9jsnPvz4stg_00047-ikWr0bEs_Ac_00005.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9894 + }, + { + "path": "iuwBKfUerlk_00027-tn7Roknj3EI_00018.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9895 + }, + { + "path": "LaHnNW6Pg10_00034-jnWv-3_xXOA_00001.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9896 + }, + { + "path": "3Fu5_tu2wrE_00016-e2oOWImMwnc_00001.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9897 + }, + { + "path": "cPk7UgRDWMM_00004-1GJo0VxRXsE_00004.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9898 + }, + { + "path": "j96kxslsnyI_00015-EVbm5SK02-E_00017.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9899 + }, + { + "path": "Uo0SYjVncmo_00003-1RNJ845WLMg_00006.wav", + "question": "The count of speakers in this spoken audio is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9900 + }, + { + "path": "aZuTqteDmAE_00006-bGw3IyqXeEw_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9901 + }, + { + "path": "ZKoj1YiIdtg_00003-_2wZVvsQYFg_00007.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9902 + }, + { + "path": "UEPsY-pWLx8_00043-UEPsY-pWLx8_00022.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9903 + }, + { + "path": "U7C8h8tebj8_00001-dg_LUd235gk_00014.wav", + "question": "What is the speaker count for this audio excerpt?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9904 + }, + { + "path": "eckQg181kQY_00009-A-Q7SmHQT70_00015.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9905 + }, + { + "path": "2nf1qU4Kfsc_00020-Puocje_yEmg_00005.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9906 + }, + { + "path": "kSECtXFNBfg_00003-Ccf97QNeRUU_00005.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9907 + }, + { + "path": "JUVui9rRBH8_00006-t4pgrDHGYaw_00002.wav", + "question": "Count the number of vocal roles in this speech.", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9908 + }, + { + "path": "SaqG27Uwf28_00019-X0KzvvCbNQk_00008.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9909 + }, + { + "path": "_KvTsHGHC00_00027-AiJinimEdKA_00002.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9910 + }, + { + "path": "mc3Twh0hTpw_00004-G5RaPcHueTI_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9911 + }, + { + "path": "Uj8oU40VEVk_00006-YHZ63r51udM_00011.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9912 + }, + { + "path": "sci80gk6bUc_00043-Rbl4PJ4Pbh0_00010.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9913 + }, + { + "path": "vqs1ZvoJ4XY_00009-A_LpQvH80KI_00001.wav", + "question": "Identify the number of people talking in this speech.", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9914 + }, + { + "path": "aaJo0S1lnhY_00005-xSPO4zmt6Fo_00002.wav", + "question": "How many different speakers convey messages in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9915 + }, + { + "path": "_6dug2ilQf0_00006-IsDHxbLR6fE_00080.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9916 + }, + { + "path": "g2ZViN4AIAE_00009-mHJGwKUVUT0_00003.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9917 + }, + { + "path": "9HBSUu7-7v0_00045-m6EmvGA_loo_00001.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9918 + }, + { + "path": "JyZY9h4z2u4_00002-QfELI4G0Fro_00014.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9919 + }, + { + "path": "8DIzctOYPgQ_00003-ShCa_0p3GoM_00002.wav", + "question": "How many people are delivering this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9920 + }, + { + "path": "o2sKvA9Geec_00002-I5ZloXyoJeU_00027.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9921 + }, + { + "path": "Yx-nufPjgVc_00023-Adpb9Dn0PoM_00046.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9922 + }, + { + "path": "xJwlycjGpSM_00001-Mm326dbpKYQ_00003.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9923 + }, + { + "path": "M5lYNe7y2QY_00017-RReqnuXX2Is_00002.wav", + "question": "Determine the total number of speakers in this audio clip.", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9924 + }, + { + "path": "HUXcMwXqdEY_00005-DKHa5phNSe8_00003.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9925 + }, + { + "path": "l9Taj6YoMBE_00003-GS5BrDTDwH4_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9926 + }, + { + "path": "Qq_ncaVFAMs_00065-Qq_ncaVFAMs_00043.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9927 + }, + { + "path": "VO520qQCt1I_00017-NtHTKtpuXT0_00034.wav", + "question": "The speech captures how many different voices?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9928 + }, + { + "path": "knMR3cAUUP4_00014-fd7HgBkKJF4_00003.wav", + "question": "Identify the quantity of distinct voices in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9929 + }, + { + "path": "6pjKcuVoHLM_00007-4PzEhoOxJAU_00007.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9930 + }, + { + "path": "6f7xJ9vFSNo_00012-efKQULo-g3Y_00001.wav", + "question": "What count of speakers is heard in this speech?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9931 + }, + { + "path": "WRgjaYn27bQ_00018-LMl5EyjpQgc_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "3", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9932 + }, + { + "path": "m6WYDkm7u_Y_00013-SaqG27Uwf28_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9933 + }, + { + "path": "G2QZRjUB_VM_00188-G2QZRjUB_VM_00148.wav", + "question": "How many speakers does this speech recording contain?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9934 + }, + { + "path": "4aJE1wFmLZE_00002-ZLzkvnq0JxI_00001.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "2", + "choice_b": "3", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9935 + }, + { + "path": "J99AzzFDZnQ_00014-J99AzzFDZnQ_00011.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9936 + }, + { + "path": "5r0dWxy17C8_00015-ukwyOYSdAlQ_00002.wav", + "question": "How many voices contribute to this audio recording of the speech?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9937 + }, + { + "path": "aMfPWRLKR0g_00042-2TlHPLqzIok_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9938 + }, + { + "path": "ecZf_nvdKvM_00003-YKRaWS6JVak_00001.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9939 + }, + { + "path": "Q-XckKQoR5Y_00003-6YkWXmYEiTQ_00026.wav", + "question": "Gauge the number of speaking persons in this segment.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "2", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9940 + }, + { + "path": "pZZgbCIOpyY_00005-qznyGO76gRU_00013.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9941 + }, + { + "path": "VzA66aG_-Lw_00010-6POKNa7L3uM_00008.wav", + "question": "How many voices are heard in this speech?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9942 + }, + { + "path": "fNJ7ttOFOkI_00008-i2SBmqNnvRw_00009.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9943 + }, + { + "path": "VB7sZ6phd1w_00003-hcOFWeQH1TE_00019.wav", + "question": "How many speakers' voices are in this segment?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9944 + }, + { + "path": "hTiHNerRF94_00002-AgaDSqjx9Gc_00008.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9945 + }, + { + "path": "2bn9i9re2yI_00034-fcaDDXnhKW0_00010.wav", + "question": "The speech is composed of how many speakers?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9946 + }, + { + "path": "nC31EBc78w0_00009-0dK_zszRMGE_00004.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9947 + }, + { + "path": "IXrZP_6JUdA_00002-7ACSjPSJZmo_00034.wav", + "question": "How many people are conversing in this speech?", + "choice_a": "2", + "choice_b": "4", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9948 + }, + { + "path": "gapJN6to2nY_00001-bd4yG4JRIUU_00006.wav", + "question": "Confirm the number of individuals speaking in this segment.", + "choice_a": "2", + "choice_b": "3", + "choice_c": "1", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9949 + }, + { + "path": "UpmWXm7Lr00_00015-y1ZY9OM4bDc_00001.wav", + "question": "The speech features how many speakers?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9950 + }, + { + "path": "NQxx1WJcaE8_00043-unlQQFwayzo_00006.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9951 + }, + { + "path": "1gtz-CUIygI_00016-q62TMalL88Y_00002.wav", + "question": "How many orators are heard in this speech?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9952 + }, + { + "path": "MsEHXiQEdLU_00003-FhithtQQzzM_00007.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9953 + }, + { + "path": "OkQHK94RjzA_00008-IvZUrnqKZsA_00006.wav", + "question": "Identify the total speakers in this speech presentation.", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9954 + }, + { + "path": "q7xxJIiRNoM_00040-yzIXg93UOIM_00021.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9955 + }, + { + "path": "WVQYsmh0Vl8_00008-VaxjUkL42s4_00002.wav", + "question": "What is the headcount of speakers in this speaking engagement?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9956 + }, + { + "path": "5lBebhRsmsU_00011-ME4IdW_T28c_00007.wav", + "question": "Distinguish the number of speakers in this speech.", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9957 + }, + { + "path": "0XmNeUnOnlg_00001-XJq32ejsjk0_00004.wav", + "question": "Assess the number of vocal contributors in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9958 + }, + { + "path": "kHiKqkxnypg_00010-eetx5dLUARE_00005.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "3", + "choice_b": "4", + "choice_c": "2", + "choice_d": "1", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9959 + }, + { + "path": "OV97ml5tgTI_00003-EyCIwLrhPKs_00005.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "1", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9960 + }, + { + "path": "_mSMCsWY96o_00001-987UrtpOII8_00001.wav", + "question": "How many participants are there in this speaking event?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "3", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9961 + }, + { + "path": "Q0A4RUFQBJU_00006-ZIapkoTWMsE_00013.wav", + "question": "What quantity of individuals are speaking in this audio?", + "choice_a": "4", + "choice_b": "1", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9962 + }, + { + "path": "w9bqihhUvCo_00002-FZ2o9sysw_0_00007.wav", + "question": "The total count of speakers in this speech is what?", + "choice_a": "2", + "choice_b": "1", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9963 + }, + { + "path": "D0t_46qPenI_00005-HACkUneHNEw_00019.wav", + "question": "How many individuals are speaking in this speech segment?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "3", + "choice_d": "4", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9964 + }, + { + "path": "r9-0pljhZqs_00015-7DRvHqgBZgQ_00001.wav", + "question": "What is the number of speakers in this spoken content?", + "choice_a": "1", + "choice_b": "2", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9965 + }, + { + "path": "R3XAl2ALqN8_00003-ioaEcM7ZCuA_00031.wav", + "question": "How many speakers' contributions are in this recording?", + "choice_a": "3", + "choice_b": "2", + "choice_c": "4", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9966 + }, + { + "path": "XBxLJGCPjFM_00009-pL0qwd1t5vk_00004.wav", + "question": "Tally the number of voices in this speech.", + "choice_a": "2", + "choice_b": "1", + "choice_c": "4", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9967 + }, + { + "path": "p3nwMpIe3pg_00005-kK0zwttMiGE_00001.wav", + "question": "The speech includes how many distinct speaking individuals?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "3", + "choice_d": "1", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9968 + }, + { + "path": "CfVIwjPbuY4_00023-EenPzVpJ6R0_00001.wav", + "question": "How many speakers make up this speech delivery?", + "choice_a": "4", + "choice_b": "3", + "choice_c": "1", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9969 + }, + { + "path": "QycjLswNm8w_00002-VJ3EMTDuu-M_00014.wav", + "question": "What number of speakers are involved in this spoken piece?", + "choice_a": "3", + "choice_b": "1", + "choice_c": "4", + "choice_d": "2", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9970 + }, + { + "path": "wk5rnBUUOhg_00010-3Xt_-rRgQu8_00008.wav", + "question": "The speech captures how many different voices?", + "choice_a": "4", + "choice_b": "2", + "choice_c": "1", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9971 + }, + { + "path": "2qTqvrCPe0I_00003-fEKabIQYh-Q_00001.wav", + "question": "The number of voices in this speech is what?", + "choice_a": "1", + "choice_b": "4", + "choice_c": "2", + "choice_d": "3", + "answer_gt": "2", + "task_name": "Speaker_Number_Verification", + "dataset_name": "voxceleb1", + "uniq_id": 9972 + }, + { + "path": "fake_20.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9973 + }, + { + "path": "fake_34.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9974 + }, + { + "path": "real_164.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9975 + }, + { + "path": "real_27.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9976 + }, + { + "path": "real_33.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9977 + }, + { + "path": "real_170.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9978 + }, + { + "path": "real_158.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9979 + }, + { + "path": "fake_233.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9980 + }, + { + "path": "real_399.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9981 + }, + { + "path": "fake_227.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9982 + }, + { + "path": "real_400.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9983 + }, + { + "path": "real_366.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9984 + }, + { + "path": "real_372.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9985 + }, + { + "path": "real_414.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9986 + }, + { + "path": "real_428.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9987 + }, + { + "path": "fake_409.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9988 + }, + { + "path": "fake_421.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9989 + }, + { + "path": "fake_347.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9990 + }, + { + "path": "fake_353.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9991 + }, + { + "path": "fake_435.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9992 + }, + { + "path": "real_212.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9993 + }, + { + "path": "real_206.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9994 + }, + { + "path": "fake_384.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9995 + }, + { + "path": "fake_390.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9996 + }, + { + "path": "fake_179.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9997 + }, + { + "path": "fake_145.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9998 + }, + { + "path": "fake_151.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 9999 + }, + { + "path": "fake_186.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10000 + }, + { + "path": "fake_192.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10001 + }, + { + "path": "fake_193.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10002 + }, + { + "path": "fake_187.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10003 + }, + { + "path": "fake_150.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10004 + }, + { + "path": "fake_144.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10005 + }, + { + "path": "fake_178.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10006 + }, + { + "path": "fake_391.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10007 + }, + { + "path": "fake_385.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10008 + }, + { + "path": "real_207.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10009 + }, + { + "path": "real_213.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10010 + }, + { + "path": "fake_352.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10011 + }, + { + "path": "fake_434.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10012 + }, + { + "path": "fake_420.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10013 + }, + { + "path": "fake_346.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10014 + }, + { + "path": "fake_408.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10015 + }, + { + "path": "real_429.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10016 + }, + { + "path": "real_373.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10017 + }, + { + "path": "real_415.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10018 + }, + { + "path": "real_401.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10019 + }, + { + "path": "real_367.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10020 + }, + { + "path": "fake_226.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10021 + }, + { + "path": "real_398.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10022 + }, + { + "path": "fake_232.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10023 + }, + { + "path": "real_159.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10024 + }, + { + "path": "real_32.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10025 + }, + { + "path": "real_171.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10026 + }, + { + "path": "real_165.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10027 + }, + { + "path": "real_26.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10028 + }, + { + "path": "fake_35.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10029 + }, + { + "path": "fake_21.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10030 + }, + { + "path": "fake_37.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10031 + }, + { + "path": "fake_23.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10032 + }, + { + "path": "real_198.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10033 + }, + { + "path": "real_173.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10034 + }, + { + "path": "real_30.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10035 + }, + { + "path": "real_24.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10036 + }, + { + "path": "real_167.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10037 + }, + { + "path": "real_18.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10038 + }, + { + "path": "fake_218.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10039 + }, + { + "path": "fake_224.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10040 + }, + { + "path": "fake_230.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10041 + }, + { + "path": "real_417.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10042 + }, + { + "path": "real_371.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10043 + }, + { + "path": "real_365.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10044 + }, + { + "path": "real_403.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10045 + }, + { + "path": "real_359.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10046 + }, + { + "path": "fake_8.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10047 + }, + { + "path": "fake_378.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10048 + }, + { + "path": "fake_436.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10049 + }, + { + "path": "fake_350.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10050 + }, + { + "path": "fake_344.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10051 + }, + { + "path": "fake_422.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10052 + }, + { + "path": "real_205.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10053 + }, + { + "path": "real_211.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10054 + }, + { + "path": "fake_393.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10055 + }, + { + "path": "real_239.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10056 + }, + { + "path": "fake_387.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10057 + }, + { + "path": "fake_152.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10058 + }, + { + "path": "fake_146.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10059 + }, + { + "path": "fake_191.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10060 + }, + { + "path": "fake_185.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10061 + }, + { + "path": "fake_184.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10062 + }, + { + "path": "fake_190.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10063 + }, + { + "path": "fake_147.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10064 + }, + { + "path": "fake_153.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10065 + }, + { + "path": "fake_386.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10066 + }, + { + "path": "real_238.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10067 + }, + { + "path": "fake_392.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10068 + }, + { + "path": "real_210.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10069 + }, + { + "path": "real_204.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10070 + }, + { + "path": "fake_345.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10071 + }, + { + "path": "fake_423.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10072 + }, + { + "path": "fake_437.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10073 + }, + { + "path": "fake_351.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10074 + }, + { + "path": "fake_379.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10075 + }, + { + "path": "fake_9.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10076 + }, + { + "path": "real_358.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10077 + }, + { + "path": "real_364.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10078 + }, + { + "path": "real_402.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10079 + }, + { + "path": "real_416.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10080 + }, + { + "path": "real_370.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10081 + }, + { + "path": "fake_231.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10082 + }, + { + "path": "fake_225.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10083 + }, + { + "path": "fake_219.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10084 + }, + { + "path": "real_19.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10085 + }, + { + "path": "real_25.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10086 + }, + { + "path": "real_166.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10087 + }, + { + "path": "real_172.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10088 + }, + { + "path": "real_31.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10089 + }, + { + "path": "fake_22.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10090 + }, + { + "path": "real_199.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10091 + }, + { + "path": "fake_36.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10092 + }, + { + "path": "fake_32.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10093 + }, + { + "path": "real_189.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10094 + }, + { + "path": "fake_26.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10095 + }, + { + "path": "real_35.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10096 + }, + { + "path": "real_176.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10097 + }, + { + "path": "real_162.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10098 + }, + { + "path": "real_21.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10099 + }, + { + "path": "fake_221.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10100 + }, + { + "path": "fake_235.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10101 + }, + { + "path": "fake_209.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10102 + }, + { + "path": "real_348.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10103 + }, + { + "path": "real_374.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10104 + }, + { + "path": "real_412.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10105 + }, + { + "path": "real_406.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10106 + }, + { + "path": "real_360.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10107 + }, + { + "path": "fake_355.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10108 + }, + { + "path": "fake_433.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10109 + }, + { + "path": "fake_427.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10110 + }, + { + "path": "fake_341.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10111 + }, + { + "path": "fake_369.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10112 + }, + { + "path": "fake_396.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10113 + }, + { + "path": "fake_382.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10114 + }, + { + "path": "real_228.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10115 + }, + { + "path": "real_200.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10116 + }, + { + "path": "real_214.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10117 + }, + { + "path": "fake_157.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10118 + }, + { + "path": "fake_143.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10119 + }, + { + "path": "fake_194.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10120 + }, + { + "path": "fake_180.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10121 + }, + { + "path": "fake_181.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10122 + }, + { + "path": "fake_195.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10123 + }, + { + "path": "fake_142.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10124 + }, + { + "path": "fake_156.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10125 + }, + { + "path": "real_215.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10126 + }, + { + "path": "real_201.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10127 + }, + { + "path": "real_229.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10128 + }, + { + "path": "fake_383.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10129 + }, + { + "path": "fake_397.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10130 + }, + { + "path": "fake_368.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10131 + }, + { + "path": "fake_426.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10132 + }, + { + "path": "fake_340.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10133 + }, + { + "path": "fake_354.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10134 + }, + { + "path": "fake_432.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10135 + }, + { + "path": "real_407.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10136 + }, + { + "path": "real_361.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10137 + }, + { + "path": "real_375.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10138 + }, + { + "path": "real_413.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10139 + }, + { + "path": "real_349.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10140 + }, + { + "path": "fake_208.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10141 + }, + { + "path": "fake_234.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10142 + }, + { + "path": "fake_220.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10143 + }, + { + "path": "real_163.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10144 + }, + { + "path": "real_20.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10145 + }, + { + "path": "real_34.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10146 + }, + { + "path": "real_177.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10147 + }, + { + "path": "fake_27.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10148 + }, + { + "path": "fake_33.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10149 + }, + { + "path": "real_188.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10150 + }, + { + "path": "fake_25.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10151 + }, + { + "path": "fake_31.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10152 + }, + { + "path": "fake_19.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10153 + }, + { + "path": "real_149.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10154 + }, + { + "path": "real_22.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10155 + }, + { + "path": "real_161.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10156 + }, + { + "path": "real_175.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10157 + }, + { + "path": "real_36.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10158 + }, + { + "path": "fake_236.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10159 + }, + { + "path": "fake_222.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10160 + }, + { + "path": "real_388.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10161 + }, + { + "path": "real_9.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10162 + }, + { + "path": "real_439.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10163 + }, + { + "path": "real_363.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10164 + }, + { + "path": "real_405.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10165 + }, + { + "path": "real_411.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10166 + }, + { + "path": "real_377.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10167 + }, + { + "path": "fake_342.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10168 + }, + { + "path": "fake_424.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10169 + }, + { + "path": "fake_430.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10170 + }, + { + "path": "fake_356.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10171 + }, + { + "path": "fake_418.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10172 + }, + { + "path": "fake_381.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10173 + }, + { + "path": "fake_395.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10174 + }, + { + "path": "real_217.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10175 + }, + { + "path": "real_203.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10176 + }, + { + "path": "fake_140.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10177 + }, + { + "path": "fake_154.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10178 + }, + { + "path": "fake_168.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10179 + }, + { + "path": "fake_183.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10180 + }, + { + "path": "fake_197.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10181 + }, + { + "path": "fake_196.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10182 + }, + { + "path": "fake_182.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10183 + }, + { + "path": "fake_169.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10184 + }, + { + "path": "fake_155.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10185 + }, + { + "path": "fake_141.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10186 + }, + { + "path": "real_202.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10187 + }, + { + "path": "real_216.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10188 + }, + { + "path": "fake_394.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10189 + }, + { + "path": "fake_380.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10190 + }, + { + "path": "fake_419.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10191 + }, + { + "path": "fake_431.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10192 + }, + { + "path": "fake_357.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10193 + }, + { + "path": "fake_343.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10194 + }, + { + "path": "fake_425.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10195 + }, + { + "path": "real_410.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10196 + }, + { + "path": "real_376.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10197 + }, + { + "path": "real_362.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10198 + }, + { + "path": "real_404.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10199 + }, + { + "path": "real_438.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10200 + }, + { + "path": "real_8.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10201 + }, + { + "path": "real_389.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10202 + }, + { + "path": "fake_223.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10203 + }, + { + "path": "fake_237.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10204 + }, + { + "path": "real_174.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10205 + }, + { + "path": "real_37.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10206 + }, + { + "path": "real_23.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10207 + }, + { + "path": "real_160.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10208 + }, + { + "path": "real_148.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10209 + }, + { + "path": "fake_18.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10210 + }, + { + "path": "fake_30.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10211 + }, + { + "path": "fake_24.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10212 + }, + { + "path": "real_87.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10213 + }, + { + "path": "real_93.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10214 + }, + { + "path": "fake_43.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10215 + }, + { + "path": "fake_57.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10216 + }, + { + "path": "real_44.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10217 + }, + { + "path": "real_107.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10218 + }, + { + "path": "real_113.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10219 + }, + { + "path": "real_50.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10220 + }, + { + "path": "fake_80.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10221 + }, + { + "path": "real_78.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10222 + }, + { + "path": "fake_94.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10223 + }, + { + "path": "fake_278.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10224 + }, + { + "path": "fake_250.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10225 + }, + { + "path": "real_488.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10226 + }, + { + "path": "fake_244.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10227 + }, + { + "path": "real_305.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10228 + }, + { + "path": "real_463.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10229 + }, + { + "path": "real_477.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10230 + }, + { + "path": "real_311.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10231 + }, + { + "path": "real_339.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10232 + }, + { + "path": "fake_293.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10233 + }, + { + "path": "fake_287.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10234 + }, + { + "path": "fake_318.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10235 + }, + { + "path": "fake_324.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10236 + }, + { + "path": "fake_442.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10237 + }, + { + "path": "fake_456.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10238 + }, + { + "path": "fake_330.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10239 + }, + { + "path": "real_271.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10240 + }, + { + "path": "real_265.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10241 + }, + { + "path": "fake_481.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10242 + }, + { + "path": "fake_495.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10243 + }, + { + "path": "real_259.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10244 + }, + { + "path": "fake_126.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10245 + }, + { + "path": "fake_132.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10246 + }, + { + "path": "fake_133.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10247 + }, + { + "path": "fake_127.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10248 + }, + { + "path": "fake_494.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10249 + }, + { + "path": "real_258.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10250 + }, + { + "path": "fake_480.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10251 + }, + { + "path": "real_264.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10252 + }, + { + "path": "real_270.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10253 + }, + { + "path": "fake_457.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10254 + }, + { + "path": "fake_331.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10255 + }, + { + "path": "fake_325.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10256 + }, + { + "path": "fake_443.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10257 + }, + { + "path": "fake_319.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10258 + }, + { + "path": "fake_286.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10259 + }, + { + "path": "fake_292.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10260 + }, + { + "path": "real_338.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10261 + }, + { + "path": "real_476.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10262 + }, + { + "path": "real_310.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10263 + }, + { + "path": "real_304.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10264 + }, + { + "path": "real_462.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10265 + }, + { + "path": "real_489.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10266 + }, + { + "path": "fake_245.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10267 + }, + { + "path": "fake_251.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10268 + }, + { + "path": "fake_279.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10269 + }, + { + "path": "fake_95.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10270 + }, + { + "path": "real_79.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10271 + }, + { + "path": "fake_81.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10272 + }, + { + "path": "real_112.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10273 + }, + { + "path": "real_51.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10274 + }, + { + "path": "real_45.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10275 + }, + { + "path": "real_106.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10276 + }, + { + "path": "fake_56.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10277 + }, + { + "path": "fake_42.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10278 + }, + { + "path": "real_92.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10279 + }, + { + "path": "real_86.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10280 + }, + { + "path": "real_90.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10281 + }, + { + "path": "fake_68.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10282 + }, + { + "path": "real_84.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10283 + }, + { + "path": "fake_54.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10284 + }, + { + "path": "fake_40.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10285 + }, + { + "path": "real_53.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10286 + }, + { + "path": "real_110.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10287 + }, + { + "path": "real_104.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10288 + }, + { + "path": "real_47.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10289 + }, + { + "path": "fake_97.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10290 + }, + { + "path": "real_138.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10291 + }, + { + "path": "fake_83.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10292 + }, + { + "path": "fake_247.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10293 + }, + { + "path": "fake_253.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10294 + }, + { + "path": "real_312.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10295 + }, + { + "path": "real_474.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10296 + }, + { + "path": "real_460.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10297 + }, + { + "path": "real_306.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10298 + }, + { + "path": "fake_284.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10299 + }, + { + "path": "real_448.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10300 + }, + { + "path": "fake_290.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10301 + }, + { + "path": "fake_469.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10302 + }, + { + "path": "real_299.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10303 + }, + { + "path": "fake_333.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10304 + }, + { + "path": "fake_455.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10305 + }, + { + "path": "fake_441.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10306 + }, + { + "path": "fake_327.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10307 + }, + { + "path": "real_266.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10308 + }, + { + "path": "real_272.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10309 + }, + { + "path": "fake_496.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10310 + }, + { + "path": "fake_482.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10311 + }, + { + "path": "fake_119.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10312 + }, + { + "path": "fake_131.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10313 + }, + { + "path": "fake_125.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10314 + }, + { + "path": "fake_124.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10315 + }, + { + "path": "fake_130.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10316 + }, + { + "path": "fake_118.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10317 + }, + { + "path": "fake_483.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10318 + }, + { + "path": "fake_497.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10319 + }, + { + "path": "real_273.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10320 + }, + { + "path": "real_267.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10321 + }, + { + "path": "fake_440.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10322 + }, + { + "path": "fake_326.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10323 + }, + { + "path": "fake_332.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10324 + }, + { + "path": "real_298.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10325 + }, + { + "path": "fake_454.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10326 + }, + { + "path": "fake_468.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10327 + }, + { + "path": "fake_291.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10328 + }, + { + "path": "fake_285.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10329 + }, + { + "path": "real_449.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10330 + }, + { + "path": "real_461.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10331 + }, + { + "path": "real_307.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10332 + }, + { + "path": "real_313.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10333 + }, + { + "path": "real_475.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10334 + }, + { + "path": "fake_252.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10335 + }, + { + "path": "fake_246.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10336 + }, + { + "path": "real_139.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10337 + }, + { + "path": "fake_82.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10338 + }, + { + "path": "fake_96.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10339 + }, + { + "path": "real_105.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10340 + }, + { + "path": "real_46.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10341 + }, + { + "path": "real_52.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10342 + }, + { + "path": "real_111.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10343 + }, + { + "path": "fake_41.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10344 + }, + { + "path": "fake_55.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10345 + }, + { + "path": "real_85.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10346 + }, + { + "path": "fake_69.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10347 + }, + { + "path": "real_91.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10348 + }, + { + "path": "fake_51.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10349 + }, + { + "path": "fake_45.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10350 + }, + { + "path": "real_95.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10351 + }, + { + "path": "fake_79.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10352 + }, + { + "path": "real_81.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10353 + }, + { + "path": "real_129.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10354 + }, + { + "path": "fake_92.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10355 + }, + { + "path": "fake_86.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10356 + }, + { + "path": "real_115.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10357 + }, + { + "path": "real_56.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10358 + }, + { + "path": "real_42.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10359 + }, + { + "path": "real_101.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10360 + }, + { + "path": "fake_242.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10361 + }, + { + "path": "fake_256.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10362 + }, + { + "path": "fake_281.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10363 + }, + { + "path": "fake_295.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10364 + }, + { + "path": "real_459.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10365 + }, + { + "path": "real_471.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10366 + }, + { + "path": "real_317.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10367 + }, + { + "path": "real_303.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10368 + }, + { + "path": "real_465.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10369 + }, + { + "path": "fake_450.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10370 + }, + { + "path": "fake_336.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10371 + }, + { + "path": "real_288.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10372 + }, + { + "path": "fake_322.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10373 + }, + { + "path": "fake_444.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10374 + }, + { + "path": "fake_478.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10375 + }, + { + "path": "fake_493.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10376 + }, + { + "path": "fake_487.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10377 + }, + { + "path": "real_263.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10378 + }, + { + "path": "real_277.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10379 + }, + { + "path": "fake_134.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10380 + }, + { + "path": "fake_120.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10381 + }, + { + "path": "fake_108.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10382 + }, + { + "path": "fake_109.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10383 + }, + { + "path": "fake_121.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10384 + }, + { + "path": "fake_135.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10385 + }, + { + "path": "real_276.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10386 + }, + { + "path": "real_262.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10387 + }, + { + "path": "fake_486.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10388 + }, + { + "path": "fake_492.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10389 + }, + { + "path": "fake_479.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10390 + }, + { + "path": "fake_323.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10391 + }, + { + "path": "real_289.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10392 + }, + { + "path": "fake_445.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10393 + }, + { + "path": "fake_451.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10394 + }, + { + "path": "fake_337.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10395 + }, + { + "path": "real_302.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10396 + }, + { + "path": "real_464.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10397 + }, + { + "path": "real_470.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10398 + }, + { + "path": "real_316.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10399 + }, + { + "path": "fake_294.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10400 + }, + { + "path": "real_458.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10401 + }, + { + "path": "fake_280.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10402 + }, + { + "path": "fake_257.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10403 + }, + { + "path": "fake_243.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10404 + }, + { + "path": "real_43.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10405 + }, + { + "path": "real_100.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10406 + }, + { + "path": "real_114.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10407 + }, + { + "path": "real_57.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10408 + }, + { + "path": "fake_87.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10409 + }, + { + "path": "real_128.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10410 + }, + { + "path": "fake_93.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10411 + }, + { + "path": "real_80.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10412 + }, + { + "path": "fake_78.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10413 + }, + { + "path": "real_94.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10414 + }, + { + "path": "fake_44.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10415 + }, + { + "path": "fake_50.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10416 + }, + { + "path": "fake_46.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10417 + }, + { + "path": "fake_52.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10418 + }, + { + "path": "real_82.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10419 + }, + { + "path": "real_96.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10420 + }, + { + "path": "fake_85.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10421 + }, + { + "path": "real_69.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10422 + }, + { + "path": "fake_91.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10423 + }, + { + "path": "real_102.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10424 + }, + { + "path": "real_41.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10425 + }, + { + "path": "real_55.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10426 + }, + { + "path": "real_116.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10427 + }, + { + "path": "real_499.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10428 + }, + { + "path": "fake_255.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10429 + }, + { + "path": "fake_241.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10430 + }, + { + "path": "fake_269.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10431 + }, + { + "path": "fake_296.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10432 + }, + { + "path": "real_328.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10433 + }, + { + "path": "fake_282.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10434 + }, + { + "path": "real_466.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10435 + }, + { + "path": "real_300.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10436 + }, + { + "path": "real_314.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10437 + }, + { + "path": "real_472.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10438 + }, + { + "path": "fake_447.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10439 + }, + { + "path": "fake_321.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10440 + }, + { + "path": "fake_335.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10441 + }, + { + "path": "fake_453.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10442 + }, + { + "path": "fake_309.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10443 + }, + { + "path": "fake_484.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10444 + }, + { + "path": "real_248.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10445 + }, + { + "path": "fake_490.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10446 + }, + { + "path": "real_274.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10447 + }, + { + "path": "real_260.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10448 + }, + { + "path": "fake_123.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10449 + }, + { + "path": "fake_137.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10450 + }, + { + "path": "fake_136.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10451 + }, + { + "path": "fake_122.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10452 + }, + { + "path": "real_261.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10453 + }, + { + "path": "real_275.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10454 + }, + { + "path": "fake_491.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10455 + }, + { + "path": "fake_485.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10456 + }, + { + "path": "real_249.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10457 + }, + { + "path": "fake_308.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10458 + }, + { + "path": "fake_334.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10459 + }, + { + "path": "fake_452.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10460 + }, + { + "path": "fake_446.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10461 + }, + { + "path": "fake_320.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10462 + }, + { + "path": "real_315.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10463 + }, + { + "path": "real_473.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10464 + }, + { + "path": "real_467.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10465 + }, + { + "path": "real_301.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10466 + }, + { + "path": "fake_283.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10467 + }, + { + "path": "real_329.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10468 + }, + { + "path": "fake_297.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10469 + }, + { + "path": "fake_268.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10470 + }, + { + "path": "fake_240.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10471 + }, + { + "path": "real_498.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10472 + }, + { + "path": "fake_254.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10473 + }, + { + "path": "real_54.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10474 + }, + { + "path": "real_117.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10475 + }, + { + "path": "real_103.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10476 + }, + { + "path": "real_40.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10477 + }, + { + "path": "fake_90.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10478 + }, + { + "path": "real_68.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10479 + }, + { + "path": "fake_84.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10480 + }, + { + "path": "real_97.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10481 + }, + { + "path": "real_83.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10482 + }, + { + "path": "fake_53.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10483 + }, + { + "path": "fake_47.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10484 + }, + { + "path": "fake_62.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10485 + }, + { + "path": "fake_76.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10486 + }, + { + "path": "real_65.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10487 + }, + { + "path": "real_126.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10488 + }, + { + "path": "real_132.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10489 + }, + { + "path": "real_71.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10490 + }, + { + "path": "fake_89.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10491 + }, + { + "path": "real_59.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10492 + }, + { + "path": "real_481.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10493 + }, + { + "path": "real_495.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10494 + }, + { + "path": "fake_259.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10495 + }, + { + "path": "fake_271.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10496 + }, + { + "path": "fake_265.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10497 + }, + { + "path": "real_324.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10498 + }, + { + "path": "real_442.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10499 + }, + { + "path": "real_456.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10500 + }, + { + "path": "real_330.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10501 + }, + { + "path": "real_318.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10502 + }, + { + "path": "real_293.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10503 + }, + { + "path": "fake_339.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10504 + }, + { + "path": "real_287.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10505 + }, + { + "path": "fake_305.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10506 + }, + { + "path": "fake_463.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10507 + }, + { + "path": "fake_477.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10508 + }, + { + "path": "fake_311.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10509 + }, + { + "path": "real_250.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10510 + }, + { + "path": "fake_488.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10511 + }, + { + "path": "real_244.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10512 + }, + { + "path": "real_278.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10513 + }, + { + "path": "fake_107.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10514 + }, + { + "path": "fake_113.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10515 + }, + { + "path": "fake_112.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10516 + }, + { + "path": "fake_106.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10517 + }, + { + "path": "real_279.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10518 + }, + { + "path": "fake_489.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10519 + }, + { + "path": "real_245.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10520 + }, + { + "path": "real_251.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10521 + }, + { + "path": "fake_476.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10522 + }, + { + "path": "fake_310.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10523 + }, + { + "path": "fake_304.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10524 + }, + { + "path": "fake_462.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10525 + }, + { + "path": "real_286.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10526 + }, + { + "path": "fake_338.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10527 + }, + { + "path": "real_292.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10528 + }, + { + "path": "real_319.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10529 + }, + { + "path": "real_457.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10530 + }, + { + "path": "real_331.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10531 + }, + { + "path": "real_325.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10532 + }, + { + "path": "real_443.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10533 + }, + { + "path": "fake_264.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10534 + }, + { + "path": "fake_270.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10535 + }, + { + "path": "real_494.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10536 + }, + { + "path": "fake_258.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10537 + }, + { + "path": "real_480.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10538 + }, + { + "path": "real_58.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10539 + }, + { + "path": "real_133.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10540 + }, + { + "path": "fake_88.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10541 + }, + { + "path": "real_70.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10542 + }, + { + "path": "real_64.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10543 + }, + { + "path": "real_127.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10544 + }, + { + "path": "fake_77.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10545 + }, + { + "path": "fake_63.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10546 + }, + { + "path": "fake_49.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10547 + }, + { + "path": "fake_75.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10548 + }, + { + "path": "fake_61.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10549 + }, + { + "path": "real_99.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10550 + }, + { + "path": "real_72.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10551 + }, + { + "path": "real_131.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10552 + }, + { + "path": "real_125.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10553 + }, + { + "path": "real_66.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10554 + }, + { + "path": "real_119.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10555 + }, + { + "path": "real_496.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10556 + }, + { + "path": "real_482.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10557 + }, + { + "path": "fake_266.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10558 + }, + { + "path": "fake_272.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10559 + }, + { + "path": "real_333.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10560 + }, + { + "path": "fake_299.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10561 + }, + { + "path": "real_455.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10562 + }, + { + "path": "real_441.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10563 + }, + { + "path": "real_327.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10564 + }, + { + "path": "real_469.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10565 + }, + { + "path": "real_284.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10566 + }, + { + "path": "fake_448.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10567 + }, + { + "path": "real_290.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10568 + }, + { + "path": "fake_312.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10569 + }, + { + "path": "fake_474.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10570 + }, + { + "path": "fake_460.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10571 + }, + { + "path": "fake_306.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10572 + }, + { + "path": "real_247.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10573 + }, + { + "path": "real_253.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10574 + }, + { + "path": "fake_138.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10575 + }, + { + "path": "fake_110.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10576 + }, + { + "path": "fake_104.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10577 + }, + { + "path": "fake_105.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10578 + }, + { + "path": "fake_111.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10579 + }, + { + "path": "fake_139.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10580 + }, + { + "path": "real_252.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10581 + }, + { + "path": "real_246.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10582 + }, + { + "path": "fake_461.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10583 + }, + { + "path": "fake_307.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10584 + }, + { + "path": "fake_313.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10585 + }, + { + "path": "fake_475.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10586 + }, + { + "path": "real_291.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10587 + }, + { + "path": "real_285.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10588 + }, + { + "path": "fake_449.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10589 + }, + { + "path": "real_468.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10590 + }, + { + "path": "real_440.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10591 + }, + { + "path": "real_326.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10592 + }, + { + "path": "fake_298.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10593 + }, + { + "path": "real_332.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10594 + }, + { + "path": "real_454.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10595 + }, + { + "path": "fake_273.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10596 + }, + { + "path": "fake_267.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10597 + }, + { + "path": "real_483.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10598 + }, + { + "path": "real_497.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10599 + }, + { + "path": "real_118.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10600 + }, + { + "path": "real_124.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10601 + }, + { + "path": "real_67.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10602 + }, + { + "path": "real_73.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10603 + }, + { + "path": "real_130.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10604 + }, + { + "path": "real_98.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10605 + }, + { + "path": "fake_60.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10606 + }, + { + "path": "fake_74.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10607 + }, + { + "path": "fake_48.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10608 + }, + { + "path": "real_88.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10609 + }, + { + "path": "fake_70.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10610 + }, + { + "path": "fake_64.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10611 + }, + { + "path": "fake_58.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10612 + }, + { + "path": "real_108.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10613 + }, + { + "path": "real_134.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10614 + }, + { + "path": "real_77.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10615 + }, + { + "path": "real_63.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10616 + }, + { + "path": "real_120.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10617 + }, + { + "path": "fake_263.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10618 + }, + { + "path": "fake_277.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10619 + }, + { + "path": "real_493.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10620 + }, + { + "path": "real_487.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10621 + }, + { + "path": "real_478.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10622 + }, + { + "path": "real_450.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10623 + }, + { + "path": "real_336.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10624 + }, + { + "path": "real_322.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10625 + }, + { + "path": "fake_288.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10626 + }, + { + "path": "real_444.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10627 + }, + { + "path": "fake_471.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10628 + }, + { + "path": "fake_317.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10629 + }, + { + "path": "fake_303.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10630 + }, + { + "path": "fake_465.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10631 + }, + { + "path": "real_281.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10632 + }, + { + "path": "real_295.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10633 + }, + { + "path": "fake_459.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10634 + }, + { + "path": "real_242.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10635 + }, + { + "path": "real_256.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10636 + }, + { + "path": "fake_115.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10637 + }, + { + "path": "fake_101.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10638 + }, + { + "path": "fake_129.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10639 + }, + { + "path": "fake_128.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10640 + }, + { + "path": "fake_100.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10641 + }, + { + "path": "fake_114.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10642 + }, + { + "path": "real_257.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10643 + }, + { + "path": "real_243.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10644 + }, + { + "path": "real_294.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10645 + }, + { + "path": "fake_458.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10646 + }, + { + "path": "real_280.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10647 + }, + { + "path": "fake_302.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10648 + }, + { + "path": "fake_464.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10649 + }, + { + "path": "fake_470.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10650 + }, + { + "path": "fake_316.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10651 + }, + { + "path": "fake_289.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10652 + }, + { + "path": "real_323.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10653 + }, + { + "path": "real_445.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10654 + }, + { + "path": "real_451.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10655 + }, + { + "path": "real_337.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10656 + }, + { + "path": "real_479.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10657 + }, + { + "path": "real_486.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10658 + }, + { + "path": "real_492.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10659 + }, + { + "path": "fake_276.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10660 + }, + { + "path": "fake_262.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10661 + }, + { + "path": "real_62.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10662 + }, + { + "path": "real_121.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10663 + }, + { + "path": "real_135.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10664 + }, + { + "path": "real_76.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10665 + }, + { + "path": "real_109.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10666 + }, + { + "path": "fake_59.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10667 + }, + { + "path": "fake_65.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10668 + }, + { + "path": "fake_71.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10669 + }, + { + "path": "real_89.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10670 + }, + { + "path": "fake_67.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10671 + }, + { + "path": "fake_73.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10672 + }, + { + "path": "real_48.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10673 + }, + { + "path": "real_123.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10674 + }, + { + "path": "fake_98.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10675 + }, + { + "path": "real_60.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10676 + }, + { + "path": "real_74.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10677 + }, + { + "path": "real_137.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10678 + }, + { + "path": "fake_274.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10679 + }, + { + "path": "fake_260.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10680 + }, + { + "path": "real_484.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10681 + }, + { + "path": "fake_248.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10682 + }, + { + "path": "real_490.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10683 + }, + { + "path": "real_309.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10684 + }, + { + "path": "real_447.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10685 + }, + { + "path": "real_321.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10686 + }, + { + "path": "real_335.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10687 + }, + { + "path": "real_453.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10688 + }, + { + "path": "fake_466.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10689 + }, + { + "path": "fake_300.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10690 + }, + { + "path": "fake_314.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10691 + }, + { + "path": "fake_472.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10692 + }, + { + "path": "real_296.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10693 + }, + { + "path": "real_282.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10694 + }, + { + "path": "fake_328.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10695 + }, + { + "path": "real_269.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10696 + }, + { + "path": "fake_499.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10697 + }, + { + "path": "real_255.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10698 + }, + { + "path": "real_241.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10699 + }, + { + "path": "fake_102.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10700 + }, + { + "path": "fake_116.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10701 + }, + { + "path": "fake_117.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10702 + }, + { + "path": "fake_103.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10703 + }, + { + "path": "real_240.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10704 + }, + { + "path": "fake_498.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10705 + }, + { + "path": "real_254.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10706 + }, + { + "path": "real_268.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10707 + }, + { + "path": "fake_329.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10708 + }, + { + "path": "real_283.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10709 + }, + { + "path": "real_297.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10710 + }, + { + "path": "fake_315.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10711 + }, + { + "path": "fake_473.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10712 + }, + { + "path": "fake_467.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10713 + }, + { + "path": "fake_301.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10714 + }, + { + "path": "real_334.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10715 + }, + { + "path": "real_452.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10716 + }, + { + "path": "real_446.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10717 + }, + { + "path": "real_320.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10718 + }, + { + "path": "real_308.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10719 + }, + { + "path": "real_491.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10720 + }, + { + "path": "real_485.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10721 + }, + { + "path": "fake_249.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10722 + }, + { + "path": "fake_261.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10723 + }, + { + "path": "fake_275.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10724 + }, + { + "path": "real_75.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10725 + }, + { + "path": "real_136.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10726 + }, + { + "path": "real_122.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10727 + }, + { + "path": "real_61.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10728 + }, + { + "path": "fake_99.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10729 + }, + { + "path": "real_49.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10730 + }, + { + "path": "fake_72.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10731 + }, + { + "path": "fake_66.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10732 + }, + { + "path": "real_186.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10733 + }, + { + "path": "fake_29.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10734 + }, + { + "path": "real_192.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10735 + }, + { + "path": "fake_15.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10736 + }, + { + "path": "real_145.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10737 + }, + { + "path": "real_12.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10738 + }, + { + "path": "real_151.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10739 + }, + { + "path": "real_179.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10740 + }, + { + "path": "real_384.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10741 + }, + { + "path": "real_390.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10742 + }, + { + "path": "real_5.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10743 + }, + { + "path": "fake_212.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10744 + }, + { + "path": "fake_206.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10745 + }, + { + "path": "real_421.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10746 + }, + { + "path": "real_347.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10747 + }, + { + "path": "real_353.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10748 + }, + { + "path": "real_435.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10749 + }, + { + "path": "real_409.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10750 + }, + { + "path": "fake_428.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10751 + }, + { + "path": "fake_400.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10752 + }, + { + "path": "fake_2.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10753 + }, + { + "path": "fake_366.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10754 + }, + { + "path": "fake_372.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10755 + }, + { + "path": "fake_414.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10756 + }, + { + "path": "fake_399.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10757 + }, + { + "path": "real_233.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10758 + }, + { + "path": "real_227.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10759 + }, + { + "path": "fake_158.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10760 + }, + { + "path": "fake_164.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10761 + }, + { + "path": "fake_170.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10762 + }, + { + "path": "fake_171.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10763 + }, + { + "path": "fake_165.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10764 + }, + { + "path": "fake_159.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10765 + }, + { + "path": "real_226.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10766 + }, + { + "path": "real_232.wav", + "question": "Assess whether this speech's pronunciation is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10767 + }, + { + "path": "fake_398.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10768 + }, + { + "path": "fake_373.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10769 + }, + { + "path": "fake_415.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10770 + }, + { + "path": "fake_3.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10771 + }, + { + "path": "fake_401.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10772 + }, + { + "path": "fake_367.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10773 + }, + { + "path": "fake_429.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10774 + }, + { + "path": "real_408.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10775 + }, + { + "path": "real_352.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10776 + }, + { + "path": "real_434.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10777 + }, + { + "path": "real_420.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10778 + }, + { + "path": "real_346.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10779 + }, + { + "path": "fake_207.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10780 + }, + { + "path": "real_4.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10781 + }, + { + "path": "fake_213.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10782 + }, + { + "path": "real_391.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10783 + }, + { + "path": "real_385.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10784 + }, + { + "path": "real_178.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10785 + }, + { + "path": "real_13.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10786 + }, + { + "path": "real_150.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10787 + }, + { + "path": "real_144.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10788 + }, + { + "path": "fake_14.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10789 + }, + { + "path": "fake_28.wav", + "question": "Is the energy level of the speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10790 + }, + { + "path": "real_193.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10791 + }, + { + "path": "real_187.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10792 + }, + { + "path": "real_191.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10793 + }, + { + "path": "real_185.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10794 + }, + { + "path": "fake_16.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10795 + }, + { + "path": "real_152.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10796 + }, + { + "path": "real_11.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10797 + }, + { + "path": "real_146.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10798 + }, + { + "path": "real_39.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10799 + }, + { + "path": "fake_239.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10800 + }, + { + "path": "real_393.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10801 + }, + { + "path": "real_387.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10802 + }, + { + "path": "fake_205.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10803 + }, + { + "path": "fake_211.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10804 + }, + { + "path": "real_6.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10805 + }, + { + "path": "real_436.wav", + "question": "Is this example of speech classified as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10806 + }, + { + "path": "real_350.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10807 + }, + { + "path": "real_344.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10808 + }, + { + "path": "real_422.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10809 + }, + { + "path": "real_378.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10810 + }, + { + "path": "fake_359.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10811 + }, + { + "path": "fake_417.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10812 + }, + { + "path": "fake_371.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10813 + }, + { + "path": "fake_365.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10814 + }, + { + "path": "fake_403.wav", + "question": "Is the pitch variation in the speech characteristic of Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10815 + }, + { + "path": "fake_1.wav", + "question": "Is the quality of this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10816 + }, + { + "path": "real_224.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10817 + }, + { + "path": "real_230.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10818 + }, + { + "path": "real_218.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10819 + }, + { + "path": "fake_173.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10820 + }, + { + "path": "fake_167.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10821 + }, + { + "path": "fake_198.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10822 + }, + { + "path": "fake_199.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10823 + }, + { + "path": "fake_166.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10824 + }, + { + "path": "fake_172.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10825 + }, + { + "path": "real_219.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10826 + }, + { + "path": "real_231.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10827 + }, + { + "path": "real_225.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10828 + }, + { + "path": "fake_364.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10829 + }, + { + "path": "fake_0.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10830 + }, + { + "path": "fake_402.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10831 + }, + { + "path": "fake_416.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10832 + }, + { + "path": "fake_370.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10833 + }, + { + "path": "fake_358.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10834 + }, + { + "path": "real_379.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10835 + }, + { + "path": "real_345.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10836 + }, + { + "path": "real_423.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10837 + }, + { + "path": "real_437.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10838 + }, + { + "path": "real_351.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10839 + }, + { + "path": "fake_210.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10840 + }, + { + "path": "real_7.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10841 + }, + { + "path": "fake_204.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10842 + }, + { + "path": "real_386.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10843 + }, + { + "path": "real_392.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10844 + }, + { + "path": "fake_238.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10845 + }, + { + "path": "real_38.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10846 + }, + { + "path": "real_147.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10847 + }, + { + "path": "real_153.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10848 + }, + { + "path": "real_10.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10849 + }, + { + "path": "fake_17.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10850 + }, + { + "path": "real_184.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10851 + }, + { + "path": "real_190.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10852 + }, + { + "path": "fake_13.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10853 + }, + { + "path": "real_194.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10854 + }, + { + "path": "real_180.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10855 + }, + { + "path": "real_28.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10856 + }, + { + "path": "real_14.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10857 + }, + { + "path": "real_157.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10858 + }, + { + "path": "real_143.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10859 + }, + { + "path": "fake_200.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10860 + }, + { + "path": "real_3.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10861 + }, + { + "path": "fake_214.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10862 + }, + { + "path": "real_396.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10863 + }, + { + "path": "fake_228.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10864 + }, + { + "path": "real_382.wav", + "question": "Is the rhythm of the speech consistent with a Real or Fake source?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10865 + }, + { + "path": "real_369.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10866 + }, + { + "path": "real_355.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10867 + }, + { + "path": "real_433.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10868 + }, + { + "path": "real_427.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10869 + }, + { + "path": "real_341.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10870 + }, + { + "path": "fake_374.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10871 + }, + { + "path": "fake_412.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10872 + }, + { + "path": "fake_406.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10873 + }, + { + "path": "fake_4.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10874 + }, + { + "path": "fake_360.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10875 + }, + { + "path": "fake_348.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10876 + }, + { + "path": "real_209.wav", + "question": "Are the pauses in the speech indicative of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10877 + }, + { + "path": "real_221.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10878 + }, + { + "path": "real_235.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10879 + }, + { + "path": "fake_176.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10880 + }, + { + "path": "fake_162.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10881 + }, + { + "path": "fake_189.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10882 + }, + { + "path": "fake_188.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10883 + }, + { + "path": "fake_163.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10884 + }, + { + "path": "fake_177.wav", + "question": "Is this speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10885 + }, + { + "path": "real_234.wav", + "question": "Is the eloquence of the speech an indicator of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10886 + }, + { + "path": "real_220.wav", + "question": "Would you label the speech as Real or Fake based on its characteristics?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10887 + }, + { + "path": "real_208.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10888 + }, + { + "path": "fake_349.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10889 + }, + { + "path": "fake_5.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10890 + }, + { + "path": "fake_407.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10891 + }, + { + "path": "fake_361.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10892 + }, + { + "path": "fake_375.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10893 + }, + { + "path": "fake_413.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10894 + }, + { + "path": "real_426.wav", + "question": "Identify this speech as Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10895 + }, + { + "path": "real_340.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10896 + }, + { + "path": "real_354.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10897 + }, + { + "path": "real_432.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10898 + }, + { + "path": "real_368.wav", + "question": "Is the expression of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10899 + }, + { + "path": "real_383.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10900 + }, + { + "path": "fake_229.wav", + "question": "Is the speech tempo consistent with a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10901 + }, + { + "path": "real_397.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10902 + }, + { + "path": "real_2.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10903 + }, + { + "path": "fake_215.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10904 + }, + { + "path": "fake_201.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10905 + }, + { + "path": "real_142.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10906 + }, + { + "path": "real_15.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10907 + }, + { + "path": "real_156.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10908 + }, + { + "path": "real_29.wav", + "question": "Is the cadence of the speaker's voice Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10909 + }, + { + "path": "real_181.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10910 + }, + { + "path": "real_195.wav", + "question": "Is the speaker's accent in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10911 + }, + { + "path": "fake_12.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10912 + }, + { + "path": "fake_10.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10913 + }, + { + "path": "fake_38.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10914 + }, + { + "path": "real_183.wav", + "question": "Based on your assessment, is this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10915 + }, + { + "path": "real_197.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10916 + }, + { + "path": "real_168.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10917 + }, + { + "path": "real_140.wav", + "question": "Is the speaker's narrative style in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10918 + }, + { + "path": "real_154.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10919 + }, + { + "path": "real_17.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10920 + }, + { + "path": "fake_217.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10921 + }, + { + "path": "real_0.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10922 + }, + { + "path": "fake_203.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10923 + }, + { + "path": "real_381.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10924 + }, + { + "path": "real_395.wav", + "question": "Is the diction in this speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10925 + }, + { + "path": "real_418.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10926 + }, + { + "path": "real_342.wav", + "question": "Is the clarity of the consonants in the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10927 + }, + { + "path": "real_424.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10928 + }, + { + "path": "real_430.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10929 + }, + { + "path": "real_356.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10930 + }, + { + "path": "fake_363.wav", + "question": "Is the speech segment Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10931 + }, + { + "path": "fake_405.wav", + "question": "Is the naturalness of the speech more likely Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10932 + }, + { + "path": "fake_7.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10933 + }, + { + "path": "fake_411.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10934 + }, + { + "path": "fake_377.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10935 + }, + { + "path": "fake_439.wav", + "question": "Determine if the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10936 + }, + { + "path": "real_236.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10937 + }, + { + "path": "fake_388.wav", + "question": "Does the speech articulate in a manner that is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10938 + }, + { + "path": "real_222.wav", + "question": "Does the speech have the warmth typically associated with Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10939 + }, + { + "path": "fake_161.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10940 + }, + { + "path": "fake_175.wav", + "question": "Is the speech's volume level typical of a Real or Fake speaker?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10941 + }, + { + "path": "fake_149.wav", + "question": "Does the speaker's enunciation in this speech sound Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10942 + }, + { + "path": "fake_148.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10943 + }, + { + "path": "fake_174.wav", + "question": "Does the clarity of the speech point towards it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10944 + }, + { + "path": "fake_160.wav", + "question": "Does this speech sound Real or Fake to you?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10945 + }, + { + "path": "real_223.wav", + "question": "Would you consider the speech's modulation to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10946 + }, + { + "path": "fake_389.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10947 + }, + { + "path": "real_237.wav", + "question": "Would you classify the resonance of the speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10948 + }, + { + "path": "fake_438.wav", + "question": "Does this speech come across as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10949 + }, + { + "path": "fake_410.wav", + "question": "Is the intonation of this speech indicative of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10950 + }, + { + "path": "fake_376.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10951 + }, + { + "path": "fake_362.wav", + "question": "Does this speech exhibit characteristics that are Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10952 + }, + { + "path": "fake_6.wav", + "question": "Is the vocal expression in this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10953 + }, + { + "path": "fake_404.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10954 + }, + { + "path": "real_431.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10955 + }, + { + "path": "real_357.wav", + "question": "Are the linguistic features of the speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10956 + }, + { + "path": "real_343.wav", + "question": "Is the speech's prosody more characteristic of being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10957 + }, + { + "path": "real_425.wav", + "question": "Is the articulation of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10958 + }, + { + "path": "real_419.wav", + "question": "Would you say the tone of this speech is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10959 + }, + { + "path": "real_394.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10960 + }, + { + "path": "real_380.wav", + "question": "Is the speech delivery Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10961 + }, + { + "path": "fake_202.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10962 + }, + { + "path": "fake_216.wav", + "question": "Is the authenticity of this speech Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10963 + }, + { + "path": "real_1.wav", + "question": "Judge whether the speech is Real or Fake.", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10964 + }, + { + "path": "real_155.wav", + "question": "Does the fluency of this speech indicate it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10965 + }, + { + "path": "real_16.wav", + "question": "Is the dynamic range of the speech typical of a Real or Fake voice?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10966 + }, + { + "path": "real_141.wav", + "question": "Is the breathiness of the speech a sign of it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10967 + }, + { + "path": "real_169.wav", + "question": "Do you consider the dialogue in this segment to be Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10968 + }, + { + "path": "real_196.wav", + "question": "Does the timbre of the speech suggest it is Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10969 + }, + { + "path": "fake_39.wav", + "question": "Does the emotional expression in this speech seem Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10970 + }, + { + "path": "real_182.wav", + "question": "Would you categorize this speech as Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "real", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10971 + }, + { + "path": "fake_11.wav", + "question": "Is the vocal quality of this speech consistent with it being Real or Fake?", + "choice_a": "fake", + "choice_b": "real", + "answer_gt": "fake", + "task_name": "Synthesized_Voice_Detection", + "dataset_name": "fake_or_real", + "uniq_id": 10972 + }, + { + "path": "Yy1saVTXsKwc.wav", + "question": "What are the exact times when 'a woman briefly talks' is present in the clip?", + "choice_a": "[0.44, 2.38]", + "choice_b": "[3.85, 10.02]", + "choice_c": "[3.42, 4.11]", + "choice_d": "[1.51, 7.99]", + "answer_gt": "[3.42, 4.11]", + "answer_gt_word": "a woman briefly talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10973 + }, + { + "path": "YwwyfGO2J__4.wav", + "question": "At which seconds does 'people speak' fully register in the sound?", + "choice_a": "[2.79, 10.01]", + "choice_b": "[0.01, 2.77]", + "choice_c": "[2.52, 6.15]", + "choice_d": "[4.54, 9.36]", + "answer_gt": "[2.79, 10.01]", + "answer_gt_word": "people speak", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10974 + }, + { + "path": "Y3kbHcjynq7k.wav", + "question": "What are the boundaries in seconds for each 'several dogs bark' presence?", + "choice_a": "[0.0, 0.72]", + "choice_b": "[1.24, 4.47]", + "choice_c": "[5.33, 8.76]", + "choice_d": "[0.0, 5.37]", + "answer_gt": "[0.0, 5.37]", + "answer_gt_word": "several dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10975 + }, + { + "path": "Yusasw640oaU.wav", + "question": "List the specific seconds of 'another man talking' appearances.", + "choice_a": "[7.88, 8.96]", + "choice_b": "[0.01, 0.39]", + "choice_c": "[1.61, 3.52]", + "choice_d": "[7.9, 8.95]", + "answer_gt": "[7.88, 8.96]", + "answer_gt_word": "another man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10976 + }, + { + "path": "Y9XtBx5HwlOc.wav", + "question": "What are the precise second intervals of 'several emergency sirens sounding' in this clip?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.1, 2.96]", + "choice_c": "[1.28, 4.09]", + "choice_d": "[2.21, 6.21]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "several emergency sirens sounding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10977 + }, + { + "path": "YtznoShjZhyc.wav", + "question": "Where is 'someone hums' consistently heard throughout the sound?", + "choice_a": "[7.8, 10.0]", + "choice_b": "[3.19, 4.41]", + "choice_c": "[7.44, 7.72]", + "choice_d": "[8.78, 9.97]", + "answer_gt": "[7.8, 10.0]", + "answer_gt_word": "someone hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10978 + }, + { + "path": "Y-67q0yYR0zk.wav", + "question": "What time segments showcase 'people in a small crowd are speaking' in its entirety?", + "choice_a": "[0.02, 9.18]", + "choice_b": "[2.51, 2.82]", + "choice_c": "[3.82, 8.46]", + "choice_d": "[5.74, 8.53]", + "answer_gt": "[0.02, 9.18]", + "answer_gt_word": "people in a small crowd are speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10979 + }, + { + "path": "Y8z8oy5mkdt4.wav", + "question": "What are the precise second intervals of 'a woman speaking' in this clip?", + "choice_a": "[0.13, 2.13]", + "choice_b": "[4.62, 5.69]", + "choice_c": "[7.14, 8.55]", + "choice_d": "[4.76, 5.72]", + "answer_gt": "[4.76, 5.72]", + "answer_gt_word": "a woman speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10980 + }, + { + "path": "Y0oKbg1lSPuU.wav", + "question": "What are the distinct time ranges of 'heavy winds blow against it' within the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.38, 7.81]", + "choice_c": "[2.23, 9.38]", + "choice_d": "[4.32, 9.57]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy winds blow against it", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10981 + }, + { + "path": "Y6xpurMt7wo4.wav", + "question": "List the specific seconds of 'heavy winds blow against it' appearances.", + "choice_a": "[0.32, 1.92]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[4.74, 5.86]", + "choice_d": "[7.52, 8.64]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy winds blow against it", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10982 + }, + { + "path": "Yy3alb754Sis.wav", + "question": "What are the comprehensive segments of 'a dog whimpers continuously' in time?", + "choice_a": "[0.0, 10.01]", + "choice_b": "[1.52, 5.94]", + "choice_c": "[1.93, 6.81]", + "choice_d": "[2.98, 7.58]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "a dog whimpers continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10983 + }, + { + "path": "Y5HRNrvHJ5KY.wav", + "question": "In what time spans is 'a man 's voice' completely audible in this recording?", + "choice_a": "[2.75, 3.15]", + "choice_b": "[3.52, 4.12]", + "choice_c": "[0.0, 8.84]", + "choice_d": "[5.95, 7.97]", + "answer_gt": "[0.0, 8.84]", + "answer_gt_word": "a man 's voice", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10984 + }, + { + "path": "Y-a9HMUYHpSo.wav", + "question": "How many seconds does each full presence of 'sounds of noisy road traffic' cover?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.28, 5.2]", + "choice_c": "[2.85, 6.42]", + "choice_d": "[3.98, 8.9]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sounds of noisy road traffic", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10985 + }, + { + "path": "YD6Xt-fOZy2k.wav", + "question": "What are the full time segments when 'sounds of noisy road traffic' appears in this sound?", + "choice_a": "[0.0, 2.99]", + "choice_b": "[9.42, 10.01]", + "choice_c": "[5.54, 6.25]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sounds of noisy road traffic", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10986 + }, + { + "path": "YlZ7B2De3kqs.wav", + "question": "What are the distinct time ranges of 'someone is heard speaking' within the audio?", + "choice_a": "[0.01, 1.04]", + "choice_b": "[7.36, 9.36]", + "choice_c": "[1.05, 10.02]", + "choice_d": "[1.77, 7.76]", + "answer_gt": "[0.01, 1.04]", + "answer_gt_word": "someone is heard speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10987 + }, + { + "path": "Y6NpPjovJ9j8.wav", + "question": "What are the comprehensive segments of 'someone is heard speaking' in time?", + "choice_a": "[0.04, 1.07]", + "choice_b": "[2.45, 3.03]", + "choice_c": "[4.33, 4.53]", + "choice_d": "[0.01, 1.04]", + "answer_gt": "[0.01, 1.04]", + "answer_gt_word": "someone is heard speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10988 + }, + { + "path": "Y0yxEvdnimGg.wav", + "question": "Identify the periods when 'birds chirp' fully resonates in the sound.", + "choice_a": "[7.35, 10.0]", + "choice_b": "[1.42, 1.65]", + "choice_c": "[3.64, 5.12]", + "choice_d": "[8.48, 8.78]", + "answer_gt": "[7.35, 10.0]", + "answer_gt_word": "birds chirp", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10989 + }, + { + "path": "Y6KINFsRbXYo.wav", + "question": "How many seconds does each full presence of 'birds chirp' cover?", + "choice_a": "[0.05, 0.34]", + "choice_b": "[0.77, 1.02]", + "choice_c": "[2.34, 2.9]", + "choice_d": "[7.35, 10.0]", + "answer_gt": "[7.35, 10.0]", + "answer_gt_word": "birds chirp", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10990 + }, + { + "path": "Y6YX6esYQEDY.wav", + "question": "What are the comprehensive segments of 'a dog pants' in time?", + "choice_a": "[0.0, 8.78]", + "choice_b": "[7.78, 8.75]", + "choice_c": "[1.17, 8.23]", + "choice_d": "[4.61, 9.2]", + "answer_gt": "[0.0, 8.78]", + "answer_gt_word": "a dog pants", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10991 + }, + { + "path": "YA1iwaCZDWE4.wav", + "question": "In what segments can 'tires squeal' be heard from start to finish?", + "choice_a": "[3.6, 7.38]", + "choice_b": "[0.0, 0.73]", + "choice_c": "[1.72, 2.34]", + "choice_d": "[8.06, 8.96]", + "answer_gt": "[3.6, 7.38]", + "answer_gt_word": "tires squeal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10992 + }, + { + "path": "YHqnSyliKTKA.wav", + "question": "Determine the full durations of 'a lawnmower engine runs in the background' in the audio file.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[7.32, 8.98]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[2.74, 9.02]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lawnmower engine runs in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10993 + }, + { + "path": "Y0FeuJL-r8m4.wav", + "question": "What time segments delineate the presence of 'man saying yep'?", + "choice_a": "[5.67, 5.99]", + "choice_b": "[0.0, 9.12]", + "choice_c": "[0.3, 6.56]", + "choice_d": "[0.44, 7.89]", + "answer_gt": "[5.67, 5.99]", + "answer_gt_word": "man saying yep", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10994 + }, + { + "path": "Y6a97tMZJNEA.wav", + "question": "Capture the timeline where 'a child is speaking' is fully included in the sound.", + "choice_a": "[0.82, 1.73]", + "choice_b": "[2.34, 3.16]", + "choice_c": "[5.04, 7.88]", + "choice_d": "[0.3, 0.86]", + "answer_gt": "[0.82, 1.73]", + "answer_gt_word": "a child is speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10995 + }, + { + "path": "YdHce3m03dgk.wav", + "question": "How many seconds does 'a person yelling' span in each occurrence?", + "choice_a": "[2.94, 3.93]", + "choice_b": "[0.0, 7.41]", + "choice_c": "[4.01, 5.08]", + "choice_d": "[4.33, 8.88]", + "answer_gt": "[2.94, 3.93]", + "answer_gt_word": "a person yelling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10996 + }, + { + "path": "YzM4rnZcNY8U.wav", + "question": "Locate every segment where 'a woman gives a speech' is completely audible.", + "choice_a": "[3.5, 5.04]", + "choice_b": "[9.26, 10.01]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[0.82, 2.35]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a woman gives a speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10997 + }, + { + "path": "Y-oy0BkpMGAk.wav", + "question": "How many seconds does each full presence of 'squealing tires' cover?", + "choice_a": "[3.93, 9.98]", + "choice_b": "[0.11, 6.37]", + "choice_c": "[2.6, 7.54]", + "choice_d": "[2.66, 9.62]", + "answer_gt": "[3.93, 9.98]", + "answer_gt_word": "squealing tires", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10998 + }, + { + "path": "YR79kDKN1-to.wav", + "question": "At which seconds does 'a horn is triggered' fully register in the sound?", + "choice_a": "[4.41, 10.02]", + "choice_b": "[0.76, 8.52]", + "choice_c": "[4.3, 8.9]", + "choice_d": "[8.1, 9.97]", + "answer_gt": "[4.41, 10.02]", + "answer_gt_word": "a horn is triggered", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 10999 + }, + { + "path": "Y2otvlKMxOQ4.wav", + "question": "How many seconds does 'a car engine revs up multiple times' span in each occurrence?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.17, 7.44]", + "choice_c": "[4.42, 7.46]", + "choice_d": "[5.0, 7.51]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a car engine revs up multiple times", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11000 + }, + { + "path": "Y5Iz0yNdnUWE.wav", + "question": "List the specific seconds of 'car horns beep' appearances.", + "choice_a": "[2.12, 3.59]", + "choice_b": "[7.37, 10.02]", + "choice_c": "[1.12, 3.68]", + "choice_d": "[3.38, 5.46]", + "answer_gt": "[2.12, 3.59]", + "answer_gt_word": "car horns beep", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11001 + }, + { + "path": "Y0N8udwD-wXE.wav", + "question": "What are the distinct time ranges of 'a sawing sound created by a power tool cutting through wood' within the audio?", + "choice_a": "[4.9, 10.02]", + "choice_b": "[0.01, 0.81]", + "choice_c": "[2.01, 2.93]", + "choice_d": "[0.36, 6.02]", + "answer_gt": "[4.9, 10.02]", + "answer_gt_word": "a sawing sound created by a power tool cutting through wood", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11002 + }, + { + "path": "YOY5503vDeA8.wav", + "question": "What ranges in seconds capture 'dogs barking' in full?", + "choice_a": "[3.83, 4.62]", + "choice_b": "[0.02, 1.2]", + "choice_c": "[2.13, 2.75]", + "choice_d": "[5.83, 6.37]", + "answer_gt": "[3.83, 4.62]", + "answer_gt_word": "dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11003 + }, + { + "path": "Yi9RHYhT0y0I.wav", + "question": "At what time points is 'engine starting loud horn' fully present in the recording?", + "choice_a": "[1.8, 2.3]", + "choice_b": "[3.09, 4.53]", + "choice_c": "[1.73, 5.75]", + "choice_d": "[5.37, 7.24]", + "answer_gt": "[1.8, 2.3]", + "answer_gt_word": "engine starting loud horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11004 + }, + { + "path": "Y4kawUPzXAlM.wav", + "question": "What are the distinct time ranges of 'people are talking' within the audio?", + "choice_a": "[5.7, 6.0]", + "choice_b": "[7.0, 9.1]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[3.39, 8.16]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11005 + }, + { + "path": "YB4FuuAYbE60.wav", + "question": "At what time intervals does 'engine revving' occur in the sound file?", + "choice_a": "[0.0, 7.52]", + "choice_b": "[4.39, 6.56]", + "choice_c": "[2.06, 4.89]", + "choice_d": "[3.61, 9.18]", + "answer_gt": "[0.0, 7.52]", + "answer_gt_word": "engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11006 + }, + { + "path": "Y5seNyvFJ_X4.wav", + "question": "How are the complete segments of 'a train horn blows' timed in this audio?", + "choice_a": "[0.0, 1.47]", + "choice_b": "[1.14, 4.26]", + "choice_c": "[1.93, 4.51]", + "choice_d": "[3.21, 8.3]", + "answer_gt": "[0.0, 1.47]", + "answer_gt_word": "a train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11007 + }, + { + "path": "Y4bp-AKTPBYE.wav", + "question": "Determine the full durations of 'an engine is idling' in the audio file.", + "choice_a": "[0.0, 9.2]", + "choice_b": "[1.0, 4.94]", + "choice_c": "[1.05, 7.18]", + "choice_d": "[3.4, 9.47]", + "answer_gt": "[0.0, 9.2]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11008 + }, + { + "path": "YD2Xc_jZllDY.wav", + "question": "Define the time intervals that fully contain 'an engine is idling' in this sound.", + "choice_a": "[0.0, 9.2]", + "choice_b": "[3.74, 6.04]", + "choice_c": "[2.38, 6.37]", + "choice_d": "[6.88, 8.77]", + "answer_gt": "[0.0, 9.2]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11009 + }, + { + "path": "Y6jrffMFpYpE.wav", + "question": "Locate every segment where 'a dog bark' is completely audible.", + "choice_a": "[2.95, 3.72]", + "choice_b": "[0.61, 3.5]", + "choice_c": "[0.85, 5.66]", + "choice_d": "[3.06, 8.25]", + "answer_gt": "[2.95, 3.72]", + "answer_gt_word": "a dog bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11010 + }, + { + "path": "YmK-KS9OfG0s.wav", + "question": "How long in seconds does 'humming' last from start to end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.08, 2.06]", + "choice_c": "[2.56, 3.2]", + "choice_d": "[8.37, 9.01]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11011 + }, + { + "path": "Y6qD9LHCKbJg.wav", + "question": "In what time spans is 'people speak' completely audible in this recording?", + "choice_a": "[2.97, 3.9]", + "choice_b": "[0.12, 10.02]", + "choice_c": "[0.92, 8.56]", + "choice_d": "[7.13, 9.7]", + "answer_gt": "[2.97, 3.9]", + "answer_gt_word": "people speak", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11012 + }, + { + "path": "Y3LeU8LHBvnU.wav", + "question": "At what time intervals does 'phone ringing' occur in the sound file?", + "choice_a": "[0.01, 6.96]", + "choice_b": "[9.52, 10.0]", + "choice_c": "[7.83, 9.48]", + "choice_d": "[1.99, 2.84]", + "answer_gt": "[7.83, 9.48]", + "answer_gt_word": "phone ringing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11013 + }, + { + "path": "Y1qWIWqV1gyI.wav", + "question": "How long in seconds does 'several dogs are whimpering' last from start to end?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.4, 2.67]", + "choice_c": "[0.86, 3.2]", + "choice_d": "[2.5, 6.73]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "several dogs are whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11014 + }, + { + "path": "Y4pGvBK7Z9t8.wav", + "question": "Where does 'a dog barks' appear from beginning to end in the sound?", + "choice_a": "[7.41, 10.01]", + "choice_b": "[0.0, 0.29]", + "choice_c": "[2.62, 7.08]", + "choice_d": "[8.57, 8.91]", + "answer_gt": "[7.41, 10.01]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11015 + }, + { + "path": "YB10u_JR3sdQ.wav", + "question": "What are the distinct time ranges of 'engines hum' within the audio?", + "choice_a": "[0.01, 10.02]", + "choice_b": "[0.01, 10.02]", + "choice_c": "[1.72, 2.82]", + "choice_d": "[1.92, 6.53]", + "answer_gt": "[0.01, 10.02]", + "answer_gt_word": "engines hum", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11016 + }, + { + "path": "Y5wN_Odph5gg.wav", + "question": "What are the comprehensive segments of 'train horn signaling' in time?", + "choice_a": "[6.26, 8.69]", + "choice_b": "[0.25, 6.59]", + "choice_c": "[2.39, 6.91]", + "choice_d": "[6.29, 8.42]", + "answer_gt": "[6.26, 8.69]", + "answer_gt_word": "train horn signaling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11017 + }, + { + "path": "Y4QHVLFkpiAI.wav", + "question": "What are the full time segments when 'a train passing by blaring its horn' appears in this sound?", + "choice_a": "[5.38, 9.8]", + "choice_b": "[0.61, 6.82]", + "choice_c": "[0.94, 6.83]", + "choice_d": "[2.03, 8.22]", + "answer_gt": "[5.38, 9.8]", + "answer_gt_word": "a train passing by blaring its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11018 + }, + { + "path": "Y1AjYKvdjOJM.wav", + "question": "What second marks correspond to the entire span of 'a train passing by blaring its horn'?", + "choice_a": "[5.38, 9.8]", + "choice_b": "[3.57, 5.1]", + "choice_c": "[1.5, 3.63]", + "choice_d": "[8.24, 9.45]", + "answer_gt": "[5.38, 9.8]", + "answer_gt_word": "a train passing by blaring its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11019 + }, + { + "path": "YaoK7MLknLCk.wav", + "question": "What are the full time segments when 'wind blows' appears in this sound?", + "choice_a": "[0.86, 1.25]", + "choice_b": "[1.73, 3.02]", + "choice_c": "[3.66, 4.31]", + "choice_d": "[4.28, 10.0]", + "answer_gt": "[4.28, 10.0]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11020 + }, + { + "path": "Y6F9yEqnd6cs.wav", + "question": "How many seconds does 'wind blows' span in each occurrence?", + "choice_a": "[1.18, 1.5]", + "choice_b": "[3.0, 3.36]", + "choice_c": "[4.28, 10.0]", + "choice_d": "[1.66, 2.57]", + "answer_gt": "[4.28, 10.0]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11021 + }, + { + "path": "Y6i5StsTUj7k.wav", + "question": "At what seconds do the full iterations of 'several people are talking loudly' start and end?", + "choice_a": "[0.07, 0.67]", + "choice_b": "[3.64, 4.86]", + "choice_c": "[8.29, 8.59]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "several people are talking loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11022 + }, + { + "path": "Y32ptEe--IM0.wav", + "question": "In what time spans is 'traffic in the distance' completely audible in this recording?", + "choice_a": "[0.01, 1.96]", + "choice_b": "[2.53, 5.68]", + "choice_c": "[0.0, 7.78]", + "choice_d": "[9.66, 10.01]", + "answer_gt": "[0.0, 7.78]", + "answer_gt_word": "traffic in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11023 + }, + { + "path": "YikhBs__x9xk.wav", + "question": "At what time points is 'a train whistle blows twice' fully present in the recording?", + "choice_a": "[6.59, 8.78]", + "choice_b": "[1.61, 6.27]", + "choice_c": "[0.0, 2.48]", + "choice_d": "[0.89, 4.74]", + "answer_gt": "[6.59, 8.78]", + "answer_gt_word": "a train whistle blows twice", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11024 + }, + { + "path": "Y0fECwjwdQUw.wav", + "question": "Enumerate the second-based occurrences of 'a car racing on a racetrack' in full.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.3, 1.15]", + "choice_c": "[1.67, 3.62]", + "choice_d": "[4.06, 7.5]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a car racing on a racetrack", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11025 + }, + { + "path": "Yx5cuQjOdM3E.wav", + "question": "What timestamps mark the full occurrences of 'a car racing on a racetrack'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.61, 2.18]", + "choice_c": "[4.29, 4.82]", + "choice_d": "[8.14, 10.02]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a car racing on a racetrack", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11026 + }, + { + "path": "Y-vAJomuQQsw.wav", + "question": "How is 'car starting but not running' distributed over the sound's duration?", + "choice_a": "[0.85, 9.14]", + "choice_b": "[0.03, 0.46]", + "choice_c": "[0.2, 1.68]", + "choice_d": "[0.3, 6.36]", + "answer_gt": "[0.85, 9.14]", + "answer_gt_word": "car starting but not running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11027 + }, + { + "path": "YyXnZjfMfzdI.wav", + "question": "Identify the occurrences of 'another man speaking' along with their time frames.", + "choice_a": "[7.97, 10.01]", + "choice_b": "[0.23, 2.2]", + "choice_c": "[3.27, 5.16]", + "choice_d": "[6.4, 7.11]", + "answer_gt": "[7.97, 10.01]", + "answer_gt_word": "another man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11028 + }, + { + "path": "Y1U1k-d1x1qw.wav", + "question": "Where does 'growling' appear from beginning to end in the sound?", + "choice_a": "[4.5, 4.8]", + "choice_b": "[1.82, 8.17]", + "choice_c": "[2.72, 9.17]", + "choice_d": "[8.0, 9.55]", + "answer_gt": "[4.5, 4.8]", + "answer_gt_word": "growling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11029 + }, + { + "path": "YJwl2o6jPXkc.wav", + "question": "Determine the full durations of 'warning bells' in the audio file.", + "choice_a": "[0.0, 3.84]", + "choice_b": "[9.38, 10.0]", + "choice_c": "[8.3, 9.43]", + "choice_d": "[1.73, 6.53]", + "answer_gt": "[8.3, 9.43]", + "answer_gt_word": "warning bells", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11030 + }, + { + "path": "YNZMDi4ZRjjQ.wav", + "question": "How long in seconds does 'warning bells' last from start to end?", + "choice_a": "[4.31, 5.9]", + "choice_b": "[8.3, 9.43]", + "choice_c": "[0.39, 1.61]", + "choice_d": "[2.57, 2.88]", + "answer_gt": "[8.3, 9.43]", + "answer_gt_word": "warning bells", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11031 + }, + { + "path": "Y34In3a1ihmA.wav", + "question": "Where is 'animals make noises in the background' consistently heard throughout the sound?", + "choice_a": "[4.72, 5.74]", + "choice_b": "[8.33, 8.54]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[1.55, 2.87]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "animals make noises in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11032 + }, + { + "path": "Y0AuWiT7RBUg.wav", + "question": "What are the precise second intervals of 'barking at him' in this clip?", + "choice_a": "[2.75, 3.07]", + "choice_b": "[2.03, 2.42]", + "choice_c": "[7.34, 7.9]", + "choice_d": "[7.46, 8.56]", + "answer_gt": "[2.75, 3.07]", + "answer_gt_word": "barking at him", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11033 + }, + { + "path": "YlJ6fEVn7IlM.wav", + "question": "Locate every segment where 'a propeller airplane is flying closer continuously' is completely audible.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[6.71, 8.38]", + "choice_c": "[0.44, 5.04]", + "choice_d": "[2.07, 9.04]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a propeller airplane is flying closer continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11034 + }, + { + "path": "Y3PPL7SS2C1Q.wav", + "question": "What are the distinct time ranges of 'a train passes on tracks' within the audio?", + "choice_a": "[1.3, 3.6]", + "choice_b": "[8.82, 9.99]", + "choice_c": "[1.27, 6.33]", + "choice_d": "[3.8, 6.74]", + "answer_gt": "[1.3, 3.6]", + "answer_gt_word": "a train passes on tracks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11035 + }, + { + "path": "YEoPFW-CgltQ.wav", + "question": "Which sections of the sound contain 'a train passes on tracks' in their entirety?", + "choice_a": "[0.0, 4.13]", + "choice_b": "[5.62, 6.64]", + "choice_c": "[8.32, 9.83]", + "choice_d": "[1.3, 3.6]", + "answer_gt": "[1.3, 3.6]", + "answer_gt_word": "a train passes on tracks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11036 + }, + { + "path": "Y3iKaVJsj6co.wav", + "question": "What are the precise second intervals of 'clapping' in this clip?", + "choice_a": "[0.01, 7.15]", + "choice_b": "[0.55, 0.83]", + "choice_c": "[1.56, 1.78]", + "choice_d": "[7.08, 7.66]", + "answer_gt": "[0.01, 7.15]", + "answer_gt_word": "clapping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11037 + }, + { + "path": "Y2z2ekcBoxRs.wav", + "question": "What time segments delineate the presence of 'clapping'?", + "choice_a": "[2.03, 2.16]", + "choice_b": "[9.7, 9.81]", + "choice_c": "[4.11, 5.94]", + "choice_d": "[0.01, 7.15]", + "answer_gt": "[0.01, 7.15]", + "answer_gt_word": "clapping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11038 + }, + { + "path": "Y83u2vOHb52Y.wav", + "question": "In what segments can 'a man talks very briefly' be heard from start to finish?", + "choice_a": "[0.52, 6.16]", + "choice_b": "[8.55, 10.02]", + "choice_c": "[7.52, 7.87]", + "choice_d": "[1.44, 2.94]", + "answer_gt": "[7.52, 7.87]", + "answer_gt_word": "a man talks very briefly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11039 + }, + { + "path": "YiIASq403bOQ.wav", + "question": "What are all the complete time segments of 'a man talks very briefly' in this sound?", + "choice_a": "[3.34, 4.91]", + "choice_b": "[7.52, 7.87]", + "choice_c": "[1.03, 2.75]", + "choice_d": "[6.8, 7.5]", + "answer_gt": "[7.52, 7.87]", + "answer_gt_word": "a man talks very briefly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11040 + }, + { + "path": "YNGPY84KXaAY.wav", + "question": "Define the time intervals that fully contain 'a man talks very briefly' in this sound.", + "choice_a": "[7.52, 7.87]", + "choice_b": "[6.34, 7.02]", + "choice_c": "[8.19, 8.43]", + "choice_d": "[0.02, 1.77]", + "answer_gt": "[7.52, 7.87]", + "answer_gt_word": "a man talks very briefly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11041 + }, + { + "path": "Y2a0pqQomO8I.wav", + "question": "Where do we hear 'more honking' in full within this sound recording?", + "choice_a": "[7.23, 10.0]", + "choice_b": "[0.04, 1.48]", + "choice_c": "[1.58, 7.09]", + "choice_d": "[4.59, 8.19]", + "answer_gt": "[7.23, 10.0]", + "answer_gt_word": "more honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11042 + }, + { + "path": "Yx5cuQjOdM3E.wav", + "question": "What are the boundaries in seconds for each 'more honking' presence?", + "choice_a": "[7.23, 10.0]", + "choice_b": "[6.66, 8.71]", + "choice_c": "[0.0, 0.31]", + "choice_d": "[1.61, 1.99]", + "answer_gt": "[7.23, 10.0]", + "answer_gt_word": "more honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11043 + }, + { + "path": "YwVi5w_NU6CM.wav", + "question": "What time segments delineate the presence of 'a ringing siren'?", + "choice_a": "[0.34, 10.02]", + "choice_b": "[4.34, 7.66]", + "choice_c": "[5.33, 7.82]", + "choice_d": "[7.36, 7.86]", + "answer_gt": "[0.34, 10.02]", + "answer_gt_word": "a ringing siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11044 + }, + { + "path": "Y-Qkeiyhg2zE.wav", + "question": "How many seconds does each full presence of 'a ringing siren' cover?", + "choice_a": "[1.34, 1.88]", + "choice_b": "[2.38, 5.84]", + "choice_c": "[0.34, 10.02]", + "choice_d": "[6.39, 10.02]", + "answer_gt": "[0.34, 10.02]", + "answer_gt_word": "a ringing siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11045 + }, + { + "path": "YpWQeV08kYR0.wav", + "question": "At what time points is 'emergency sirens going off' fully present in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.56, 1.85]", + "choice_d": "[1.23, 5.96]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "emergency sirens going off", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11046 + }, + { + "path": "Y1NrW4c8fleU.wav", + "question": "Which sections of the sound contain 'an engine running in the background' in their entirety?", + "choice_a": "[0.0, 8.86]", + "choice_b": "[0.02, 6.27]", + "choice_c": "[0.86, 8.09]", + "choice_d": "[3.54, 9.94]", + "answer_gt": "[0.0, 8.86]", + "answer_gt_word": "an engine running in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11047 + }, + { + "path": "Y5ckHyaUmfEI.wav", + "question": "What time segments showcase 'a small dog howl' in its entirety?", + "choice_a": "[1.18, 2.03]", + "choice_b": "[4.24, 4.88]", + "choice_c": "[6.23, 6.55]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a small dog howl", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11048 + }, + { + "path": "YIvBTc8y7s0M.wav", + "question": "What are the intervals in seconds of each 'engines roar' appearance?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.29, 3.33]", + "choice_c": "[0.64, 9.99]", + "choice_d": "[2.52, 9.99]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "engines roar", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11049 + }, + { + "path": "YHAjoHORA-Lg.wav", + "question": "At what time points is 'passes by' fully present in the recording?", + "choice_a": "[6.98, 10.0]", + "choice_b": "[2.99, 4.64]", + "choice_c": "[5.36, 6.81]", + "choice_d": "[0.0, 2.96]", + "answer_gt": "[6.98, 10.0]", + "answer_gt_word": "passes by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11050 + }, + { + "path": "YvhIYOLTyksQ.wav", + "question": "Locate every segment where 'a sporadic police siren' is completely audible.", + "choice_a": "[4.78, 6.78]", + "choice_b": "[0.61, 3.91]", + "choice_c": "[7.29, 8.96]", + "choice_d": "[9.57, 10.0]", + "answer_gt": "[4.78, 6.78]", + "answer_gt_word": "a sporadic police siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11051 + }, + { + "path": "Y9ucxDI35nig.wav", + "question": "What are all the complete time segments of 'a car engine hums loudly' in this sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[4.26, 5.29]", + "choice_c": "[5.88, 6.47]", + "choice_d": "[7.94, 10.0]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a car engine hums loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11052 + }, + { + "path": "Y5ndGFwBfDVY.wav", + "question": "Where is 'a car engine hums loudly' consistently heard throughout the sound?", + "choice_a": "[3.85, 4.75]", + "choice_b": "[5.82, 7.68]", + "choice_c": "[0.0, 0.58]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a car engine hums loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11053 + }, + { + "path": "Y7kYXCkEGxwI.wav", + "question": "At what time points is 'man hollering' fully present in the recording?", + "choice_a": "[1.24, 1.46]", + "choice_b": "[1.99, 2.16]", + "choice_c": "[2.77, 3.2]", + "choice_d": "[6.69, 7.92]", + "answer_gt": "[6.69, 7.92]", + "answer_gt_word": "man hollering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11054 + }, + { + "path": "Y4Ka46DJjDZk.wav", + "question": "What are the comprehensive timeframes of 'man hollering' in this clip?", + "choice_a": "[0.0, 3.54]", + "choice_b": "[6.69, 7.92]", + "choice_c": "[6.84, 7.45]", + "choice_d": "[1.45, 2.51]", + "answer_gt": "[6.69, 7.92]", + "answer_gt_word": "man hollering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11055 + }, + { + "path": "Y8DrIJDw3b_M.wav", + "question": "What ranges in seconds capture 'dogs barking' in full?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[6.62, 8.96]", + "choice_d": "[7.84, 9.33]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11056 + }, + { + "path": "Y4PcRN6DC7co.wav", + "question": "Capture the timeline where 'a man speaks' is fully included in the sound.", + "choice_a": "[0.23, 7.61]", + "choice_b": "[9.74, 10.02]", + "choice_c": "[6.31, 8.04]", + "choice_d": "[8.07, 9.46]", + "answer_gt": "[6.31, 8.04]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11057 + }, + { + "path": "Y165obDPbUVw.wav", + "question": "What time segments showcase 'whimpering animal' in its entirety?", + "choice_a": "[7.28, 7.5]", + "choice_b": "[8.11, 8.39]", + "choice_c": "[0.66, 4.28]", + "choice_d": "[3.84, 5.34]", + "answer_gt": "[0.66, 4.28]", + "answer_gt_word": "whimpering animal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11058 + }, + { + "path": "YGB-RPZWnk50.wav", + "question": "Define the time intervals that fully contain 'whimpering animal' in this sound.", + "choice_a": "[0.66, 4.28]", + "choice_b": "[4.52, 5.11]", + "choice_c": "[5.82, 6.31]", + "choice_d": "[6.9, 7.89]", + "answer_gt": "[0.66, 4.28]", + "answer_gt_word": "whimpering animal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11059 + }, + { + "path": "Y-ybHTybEnh8.wav", + "question": "What are the comprehensive timeframes of 'whimpering animal' in this clip?", + "choice_a": "[0.29, 0.66]", + "choice_b": "[2.52, 3.48]", + "choice_c": "[0.66, 4.28]", + "choice_d": "[7.58, 8.5]", + "answer_gt": "[0.66, 4.28]", + "answer_gt_word": "whimpering animal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11060 + }, + { + "path": "YgIdmmdHdAJ8.wav", + "question": "At what timestamps does the complete 'wind blowing' occur?", + "choice_a": "[1.5, 2.08]", + "choice_b": "[3.35, 4.4]", + "choice_c": "[7.64, 8.07]", + "choice_d": "[8.14, 10.01]", + "answer_gt": "[8.14, 10.01]", + "answer_gt_word": "wind blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11061 + }, + { + "path": "YsofxkNWaP0s.wav", + "question": "How long in seconds does 'an aircraft engine approaches' last from start to end?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.09, 1.63]", + "choice_c": "[4.52, 5.07]", + "choice_d": "[5.66, 6.29]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an aircraft engine approaches", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11062 + }, + { + "path": "Y1p1ekNTgCQ0.wav", + "question": "How long in seconds does 'train passes' last from start to end?", + "choice_a": "[6.4, 7.0]", + "choice_b": "[7.5, 8.6]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[2.72, 6.35]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "train passes", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11063 + }, + { + "path": "Y7UTiEhYDBQs.wav", + "question": "Where is 'animals whimper' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[1.55, 4.89]", + "choice_d": "[2.41, 7.95]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "animals whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11064 + }, + { + "path": "Y2CDH8aFgvu0.wav", + "question": "Identify the occurrences of 'animals whimper' along with their time frames.", + "choice_a": "[0.5, 1.03]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[5.84, 10.02]", + "choice_d": "[2.08, 2.83]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "animals whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11065 + }, + { + "path": "Y-1oRfe0FQI8.wav", + "question": "At what seconds do the full iterations of 'train chugging' start and end?", + "choice_a": "[0.8, 10.0]", + "choice_b": "[0.02, 1.27]", + "choice_c": "[0.95, 10.0]", + "choice_d": "[5.9, 8.44]", + "answer_gt": "[0.8, 10.0]", + "answer_gt_word": "train chugging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11066 + }, + { + "path": "Y-55UGnqZsOo.wav", + "question": "Where is 'train chugging' consistently heard throughout the sound?", + "choice_a": "[0.41, 1.51]", + "choice_b": "[3.7, 4.22]", + "choice_c": "[0.8, 10.0]", + "choice_d": "[7.72, 8.11]", + "answer_gt": "[0.8, 10.0]", + "answer_gt_word": "train chugging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11067 + }, + { + "path": "Y-5IoPSM59xM.wav", + "question": "Where does 'a small dog barking' appear from beginning to end in the sound?", + "choice_a": "[6.99, 7.2]", + "choice_b": "[1.68, 8.77]", + "choice_c": "[3.59, 8.94]", + "choice_d": "[6.79, 9.11]", + "answer_gt": "[6.99, 7.2]", + "answer_gt_word": "a small dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11068 + }, + { + "path": "Y-AlpLvWoLKs.wav", + "question": "What are the comprehensive timeframes of 'applauding' in this clip?", + "choice_a": "[0.97, 2.38]", + "choice_b": "[6.25, 10.0]", + "choice_c": "[9.17, 10.0]", + "choice_d": "[4.96, 9.63]", + "answer_gt": "[9.17, 10.0]", + "answer_gt_word": "applauding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11069 + }, + { + "path": "Y-B2i3dzzdso.wav", + "question": "What second marks correspond to the entire span of 'applauding'?", + "choice_a": "[9.17, 10.0]", + "choice_b": "[3.29, 10.0]", + "choice_c": "[0.44, 2.22]", + "choice_d": "[4.57, 5.18]", + "answer_gt": "[9.17, 10.0]", + "answer_gt_word": "applauding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11070 + }, + { + "path": "Y-HVUL5tNzdo.wav", + "question": "What time segments delineate the presence of 'applauding'?", + "choice_a": "[0.42, 1.41]", + "choice_b": "[4.39, 4.92]", + "choice_c": "[6.8, 7.64]", + "choice_d": "[9.17, 10.0]", + "answer_gt": "[9.17, 10.0]", + "answer_gt_word": "applauding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11071 + }, + { + "path": "Y-HfyfiDVqAo.wav", + "question": "In what time spans is 'applauding' completely audible in this recording?", + "choice_a": "[0.69, 1.34]", + "choice_b": "[2.91, 4.55]", + "choice_c": "[9.17, 10.0]", + "choice_d": "[7.85, 8.8]", + "answer_gt": "[9.17, 10.0]", + "answer_gt_word": "applauding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11072 + }, + { + "path": "Y-KjE40DlSdU.wav", + "question": "What are all the complete time segments of 'train chugging down the tracks' in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 4.33]", + "choice_c": "[6.27, 10.0]", + "choice_d": "[6.74, 8.1]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "train chugging down the tracks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11073 + }, + { + "path": "Y-MZGb2-2DoA.wav", + "question": "In what segments can 'automobile engine being revved' be heard from start to finish?", + "choice_a": "[0.01, 10.02]", + "choice_b": "[1.32, 3.48]", + "choice_c": "[1.57, 8.24]", + "choice_d": "[2.08, 9.43]", + "answer_gt": "[0.01, 10.02]", + "answer_gt_word": "automobile engine being revved", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11074 + }, + { + "path": "Y-VCN1giSq7c.wav", + "question": "Where does 'a barking sound is heard' appear from beginning to end in the sound?", + "choice_a": "[0.0, 1.91]", + "choice_b": "[3.65, 4.59]", + "choice_c": "[2.05, 2.91]", + "choice_d": "[0.31, 8.44]", + "answer_gt": "[2.05, 2.91]", + "answer_gt_word": "a barking sound is heard", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11075 + }, + { + "path": "Y-WE3eMj91tI.wav", + "question": "Define the time intervals that fully contain 'meow' in this sound.", + "choice_a": "[5.83, 6.24]", + "choice_b": "[5.86, 6.27]", + "choice_c": "[2.17, 2.49]", + "choice_d": "[2.99, 3.76]", + "answer_gt": "[5.83, 6.24]", + "answer_gt_word": "meow", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11076 + }, + { + "path": "Y-XWE0n4zGbU.wav", + "question": "At which seconds does 'engine sound and accelerating sound' fully register in the sound?", + "choice_a": "[0.09, 8.97]", + "choice_b": "[0.89, 8.84]", + "choice_c": "[6.06, 9.4]", + "choice_d": "[6.74, 9.74]", + "answer_gt": "[0.09, 8.97]", + "answer_gt_word": "engine sound and accelerating sound", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11077 + }, + { + "path": "Y-XZsrkx1QTg.wav", + "question": "What are all the complete time segments of 'engine sound and accelerating sound' in this sound?", + "choice_a": "[0.0, 7.02]", + "choice_b": "[0.09, 8.97]", + "choice_c": "[1.12, 2.2]", + "choice_d": "[5.35, 7.99]", + "answer_gt": "[0.09, 8.97]", + "answer_gt_word": "engine sound and accelerating sound", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11078 + }, + { + "path": "Y-YDgnw597_g.wav", + "question": "In what segments can 'the crowd clapping' be heard from start to finish?", + "choice_a": "[6.14, 9.99]", + "choice_b": "[0.94, 5.73]", + "choice_c": "[0.19, 4.42]", + "choice_d": "[0.57, 9.21]", + "answer_gt": "[6.14, 9.99]", + "answer_gt_word": "the crowd clapping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11079 + }, + { + "path": "Y-dDpGRPHzM0.wav", + "question": "At what seconds do the full iterations of 'the crowd clapping' start and end?", + "choice_a": "[6.14, 9.99]", + "choice_b": "[4.11, 5.06]", + "choice_c": "[8.23, 9.9]", + "choice_d": "[3.46, 3.75]", + "answer_gt": "[6.14, 9.99]", + "answer_gt_word": "the crowd clapping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11080 + }, + { + "path": "Y-gAYyzIl2-w.wav", + "question": "What are the comprehensive segments of 'an engine starting' in time?", + "choice_a": "[2.0, 3.03]", + "choice_b": "[2.72, 10.0]", + "choice_c": "[0.99, 9.22]", + "choice_d": "[1.39, 9.7]", + "answer_gt": "[2.0, 3.03]", + "answer_gt_word": "an engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11081 + }, + { + "path": "Y-iYTyeANFVM.wav", + "question": "At what time intervals does 'an engine starting' occur in the sound file?", + "choice_a": "[1.28, 3.18]", + "choice_b": "[2.0, 3.03]", + "choice_c": "[8.32, 10.0]", + "choice_d": "[1.28, 3.18]", + "answer_gt": "[2.0, 3.03]", + "answer_gt_word": "an engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11082 + }, + { + "path": "Y-jCrIH2sB0E.wav", + "question": "Where do we hear 'an engine starting' in full within this sound recording?", + "choice_a": "[1.79, 3.97]", + "choice_b": "[4.55, 8.3]", + "choice_c": "[5.28, 5.53]", + "choice_d": "[2.0, 3.03]", + "answer_gt": "[2.0, 3.03]", + "answer_gt_word": "an engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11083 + }, + { + "path": "Y-jCyCT3KjZA.wav", + "question": "Determine the full durations of 'tunnel sounding noise' in the audio file.", + "choice_a": "[0.86, 9.08]", + "choice_b": "[5.21, 7.69]", + "choice_c": "[5.2, 7.69]", + "choice_d": "[0.82, 3.6]", + "answer_gt": "[0.86, 9.08]", + "answer_gt_word": "tunnel sounding noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11084 + }, + { + "path": "Y-mb4Fw4Z0xg.wav", + "question": "How many seconds does each full presence of 'race cars are racing' cover?", + "choice_a": "[0.01, 9.02]", + "choice_b": "[5.44, 5.95]", + "choice_c": "[0.88, 5.9]", + "choice_d": "[5.34, 8.78]", + "answer_gt": "[0.01, 9.02]", + "answer_gt_word": "race cars are racing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11085 + }, + { + "path": "Y-pmqKob1EHs.wav", + "question": "How long does 'a vehicle trying to start its engine' last each time it appears in the audio?", + "choice_a": "[0.02, 10.01]", + "choice_b": "[0.45, 7.2]", + "choice_c": "[6.52, 8.72]", + "choice_d": "[7.14, 9.24]", + "answer_gt": "[0.02, 10.01]", + "answer_gt_word": "a vehicle trying to start its engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11086 + }, + { + "path": "Y-q3fQSq7Cns.wav", + "question": "Document the full appearances of 'bang 's' in terms of seconds.", + "choice_a": "[1.55, 7.0]", + "choice_b": "[1.6, 7.0]", + "choice_c": "[1.19, 4.12]", + "choice_d": "[3.53, 8.48]", + "answer_gt": "[1.55, 7.0]", + "answer_gt_word": "bang 's", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11087 + }, + { + "path": "Y-ttKN5Ba368.wav", + "question": "Where is 'bang 's' found in its entirety within the audio's timeline?", + "choice_a": "[0.01, 5.17]", + "choice_b": "[6.49, 10.01]", + "choice_c": "[1.55, 7.0]", + "choice_d": "[7.14, 7.33]", + "answer_gt": "[1.55, 7.0]", + "answer_gt_word": "bang 's", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11088 + }, + { + "path": "Y-zXeN78-kV4.wav", + "question": "What are the intervals in seconds of each 'a engine is running' appearance?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.48, 5.66]", + "choice_c": "[1.88, 7.67]", + "choice_d": "[3.02, 9.68]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11089 + }, + { + "path": "Y00PMurXGTMc.wav", + "question": "What are the comprehensive timeframes of 'a engine is running' in this clip?", + "choice_a": "[0.14, 7.14]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.02, 1.85]", + "choice_d": "[4.64, 5.46]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11090 + }, + { + "path": "Y06seVx9COw8.wav", + "question": "At what time points is 'cat caterwauling' fully present in the recording?", + "choice_a": "[1.15, 4.59]", + "choice_b": "[1.37, 5.22]", + "choice_c": "[1.4, 7.06]", + "choice_d": "[1.67, 9.06]", + "answer_gt": "[1.15, 4.59]", + "answer_gt_word": "cat caterwauling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11091 + }, + { + "path": "Y089a2UWj08U.wav", + "question": "List the specific seconds of 'siren is blaring' appearances.", + "choice_a": "[0.01, 10.01]", + "choice_b": "[0.0, 0.56]", + "choice_c": "[5.18, 8.2]", + "choice_d": "[8.61, 10.0]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "siren is blaring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11092 + }, + { + "path": "Y0DL92_Gd79M.wav", + "question": "Provide the start and end times of 'an adult male' within this audio.", + "choice_a": "[5.99, 6.3]", + "choice_b": "[1.77, 3.0]", + "choice_c": "[5.42, 5.99]", + "choice_d": "[6.33, 7.92]", + "answer_gt": "[5.99, 6.3]", + "answer_gt_word": "an adult male", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11093 + }, + { + "path": "Y0G-Qx3Gv01s.wav", + "question": "What are the comprehensive segments of 'a guy talking' in time?", + "choice_a": "[1.32, 3.16]", + "choice_b": "[3.88, 5.52]", + "choice_c": "[0.03, 1.4]", + "choice_d": "[2.0, 3.45]", + "answer_gt": "[1.32, 3.16]", + "answer_gt_word": "a guy talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11094 + }, + { + "path": "Y0GJCbGp_v1g.wav", + "question": "Identify the periods when 'a woman speaks' fully resonates in the sound.", + "choice_a": "[3.38, 4.34]", + "choice_b": "[6.19, 7.26]", + "choice_c": "[8.63, 9.85]", + "choice_d": "[5.31, 8.0]", + "answer_gt": "[3.38, 4.34]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11095 + }, + { + "path": "Y0IGvw7Eplrg.wav", + "question": "At which time markers does 'the idling of an engine' fully unfold in the sound?", + "choice_a": "[0.01, 9.57]", + "choice_b": "[1.94, 4.84]", + "choice_c": "[2.27, 5.01]", + "choice_d": "[3.97, 8.54]", + "answer_gt": "[0.01, 9.57]", + "answer_gt_word": "the idling of an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11096 + }, + { + "path": "Y0JqQzE4wGEo.wav", + "question": "What are the distinct time ranges of 'man speaking over dog barking' within the audio?", + "choice_a": "[0.0, 3.82]", + "choice_b": "[0.51, 5.69]", + "choice_c": "[2.65, 6.83]", + "choice_d": "[4.82, 6.95]", + "answer_gt": "[0.0, 3.82]", + "answer_gt_word": "man speaking over dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11097 + }, + { + "path": "Y0LezpTtLGik.wav", + "question": "What is the range in seconds for each full 'a motorcycle whizzing by' in the recording?", + "choice_a": "[0.0, 9.76]", + "choice_b": "[1.5, 4.64]", + "choice_c": "[1.73, 5.18]", + "choice_d": "[2.31, 7.67]", + "answer_gt": "[0.0, 9.76]", + "answer_gt_word": "a motorcycle whizzing by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11098 + }, + { + "path": "Y0PJ6PRTzoWs.wav", + "question": "List the periods when 'a train honks its horn' is completely played in the sound.", + "choice_a": "[1.65, 3.03]", + "choice_b": "[2.78, 4.45]", + "choice_c": "[3.19, 5.11]", + "choice_d": "[3.43, 5.12]", + "answer_gt": "[1.65, 3.03]", + "answer_gt_word": "a train honks its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11099 + }, + { + "path": "Y0QSg9bSOgbA.wav", + "question": "How long in seconds does 'a train honks its horn' last from start to end?", + "choice_a": "[0.4, 0.82]", + "choice_b": "[4.45, 4.93]", + "choice_c": "[1.65, 3.03]", + "choice_d": "[0.77, 1.96]", + "answer_gt": "[1.65, 3.03]", + "answer_gt_word": "a train honks its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11100 + }, + { + "path": "Y0T0MFoavHic.wav", + "question": "At what seconds do the full iterations of 'a dog barks once' start and end?", + "choice_a": "[4.62, 4.97]", + "choice_b": "[0.0, 0.83]", + "choice_c": "[2.18, 5.71]", + "choice_d": "[6.4, 10.0]", + "answer_gt": "[4.62, 4.97]", + "answer_gt_word": "a dog barks once", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11101 + }, + { + "path": "Y0TumqT-_n9k.wav", + "question": "Document the full appearances of 'loud wind' in terms of seconds.", + "choice_a": "[0.0, 9.24]", + "choice_b": "[0.0, 9.24]", + "choice_c": "[1.3, 3.92]", + "choice_d": "[2.21, 5.26]", + "answer_gt": "[0.0, 9.24]", + "answer_gt_word": "loud wind", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11102 + }, + { + "path": "Y0VlXTRYdzcU.wav", + "question": "Provide the start and end times of 'a emergency vehicle siren sounds' within this audio.", + "choice_a": "[0.01, 9.27]", + "choice_b": "[1.38, 2.33]", + "choice_c": "[2.82, 3.89]", + "choice_d": "[5.7, 6.87]", + "answer_gt": "[0.01, 9.27]", + "answer_gt_word": "a emergency vehicle siren sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11103 + }, + { + "path": "Y0dVfocjWM28.wav", + "question": "Where do we hear 'a car starting' in full within this sound recording?", + "choice_a": "[3.42, 6.9]", + "choice_b": "[0.27, 7.37]", + "choice_c": "[6.1, 8.65]", + "choice_d": "[7.24, 8.7]", + "answer_gt": "[3.42, 6.9]", + "answer_gt_word": "a car starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11104 + }, + { + "path": "Y0fEnGyWYLCc.wav", + "question": "Identify the complete instances of 'car revving its engine' within the sound timeline.", + "choice_a": "[3.81, 10.0]", + "choice_b": "[1.36, 6.71]", + "choice_c": "[4.24, 8.66]", + "choice_d": "[6.57, 9.88]", + "answer_gt": "[3.81, 10.0]", + "answer_gt_word": "car revving its engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11105 + }, + { + "path": "Y0gD6-Tes60A.wav", + "question": "What are the distinct time ranges of 'a larger bark' within the audio?", + "choice_a": "[6.01, 6.51]", + "choice_b": "[2.13, 2.56]", + "choice_c": "[3.22, 3.81]", + "choice_d": "[4.35, 4.9]", + "answer_gt": "[6.01, 6.51]", + "answer_gt_word": "a larger bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11106 + }, + { + "path": "Y0kueqaUidvk.wav", + "question": "What are the distinct time ranges of 'water is running' within the audio?", + "choice_a": "[0.03, 10.0]", + "choice_b": "[0.87, 5.55]", + "choice_c": "[8.62, 9.07]", + "choice_d": "[0.03, 10.0]", + "answer_gt": "[0.03, 10.0]", + "answer_gt_word": "water is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11107 + }, + { + "path": "Y0ojTL4MmXjQ.wav", + "question": "What are the boundaries in seconds for each 'emergency fire truck siren is revving' presence?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.98, 4.02]", + "choice_c": "[3.56, 4.15]", + "choice_d": "[3.7, 7.77]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "emergency fire truck siren is revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11108 + }, + { + "path": "Y0pNLmwA8DdI.wav", + "question": "What are the comprehensive timeframes of 'birds chirping in the background' in this clip?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.29, 7.09]", + "choice_c": "[2.04, 8.13]", + "choice_d": "[4.13, 8.35]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "birds chirping in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11109 + }, + { + "path": "Y0rcAa1-6z28.wav", + "question": "At which seconds does 'airplane propeller' fully register in the sound?", + "choice_a": "[2.1, 9.98]", + "choice_b": "[0.14, 6.84]", + "choice_c": "[2.68, 8.96]", + "choice_d": "[3.24, 9.21]", + "answer_gt": "[2.1, 9.98]", + "answer_gt_word": "airplane propeller", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11111 + }, + { + "path": "Y0uab4-3d6MM.wav", + "question": "What second marks correspond to the entire span of 'thumping is present'?", + "choice_a": "[0.0, 4.6]", + "choice_b": "[5.9, 8.33]", + "choice_c": "[9.24, 9.62]", + "choice_d": "[5.28, 5.97]", + "answer_gt": "[5.28, 5.97]", + "answer_gt_word": "thumping is present", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11112 + }, + { + "path": "Y0vOgLGk1dt0.wav", + "question": "What second marks correspond to the entire span of 'a woman speaks'?", + "choice_a": "[0.0, 3.92]", + "choice_b": "[1.97, 2.53]", + "choice_c": "[2.77, 8.06]", + "choice_d": "[3.84, 9.4]", + "answer_gt": "[0.0, 3.92]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11113 + }, + { + "path": "Y0xIwZLID2i4.wav", + "question": "What timestamps mark the full occurrences of 'racing vehicle engines are running'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.88, 8.18]", + "choice_c": "[5.18, 8.69]", + "choice_d": "[7.02, 9.26]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "racing vehicle engines are running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11114 + }, + { + "path": "Y1-jv-736aIo.wav", + "question": "Provide the start and end times of 'birds chirping' within this audio.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[4.07, 8.55]", + "choice_c": "[4.44, 8.81]", + "choice_d": "[6.33, 9.64]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "birds chirping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11116 + }, + { + "path": "Y164fV2R8E4Y.wav", + "question": "Identify the periods when 'applause' fully resonates in the sound.", + "choice_a": "[0.0, 5.83]", + "choice_b": "[5.72, 6.35]", + "choice_c": "[7.01, 10.01]", + "choice_d": "[4.27, 5.54]", + "answer_gt": "[0.0, 5.83]", + "answer_gt_word": "applause", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11117 + }, + { + "path": "Y16KqWYIlR0s.wav", + "question": "How long does 'applause' last each time it appears in the audio?", + "choice_a": "[0.0, 1.18]", + "choice_b": "[1.86, 5.91]", + "choice_c": "[0.0, 5.83]", + "choice_d": "[0.31, 1.57]", + "answer_gt": "[0.0, 5.83]", + "answer_gt_word": "applause", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11118 + }, + { + "path": "Y18FN3fDI25A.wav", + "question": "Where does 'animals growl' appear from beginning to end in the sound?", + "choice_a": "[7.71, 10.0]", + "choice_b": "[7.8, 9.43]", + "choice_c": "[9.42, 10.0]", + "choice_d": "[2.62, 8.18]", + "answer_gt": "[7.71, 10.0]", + "answer_gt_word": "animals growl", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11119 + }, + { + "path": "Y19rWZojaOK8.wav", + "question": "What are all the complete time segments of 'a lot of dogs barking' in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.7, 3.56]", + "choice_c": "[1.75, 6.31]", + "choice_d": "[2.27, 7.15]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lot of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11120 + }, + { + "path": "Y1ATOT2TZ5Gg.wav", + "question": "What are the specific start and stop times of 'a lot of dogs barking' in the sound?", + "choice_a": "[0.0, 0.83]", + "choice_b": "[1.21, 5.97]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[0.52, 1.53]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lot of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11121 + }, + { + "path": "Y1AejlUdd5gs.wav", + "question": "How are the complete segments of 'a lot of dogs barking' timed in this audio?", + "choice_a": "[0.0, 0.38]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[5.37, 5.6]", + "choice_d": "[0.0, 0.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lot of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11122 + }, + { + "path": "Y1F7JD5A9iTI.wav", + "question": "What is the range in seconds for each full 'a lot of dogs barking' in the recording?", + "choice_a": "[0.25, 2.77]", + "choice_b": "[3.52, 3.9]", + "choice_c": "[9.33, 10.02]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lot of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11123 + }, + { + "path": "Y1J-SyJKJZa0.wav", + "question": "What are the comprehensive timeframes of 'a lot of dogs barking' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.76, 4.11]", + "choice_c": "[4.56, 9.08]", + "choice_d": "[6.2, 6.48]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a lot of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11124 + }, + { + "path": "Y1LGQ1dUHT0Q.wav", + "question": "Define the time intervals that fully contain 'lots of dogs barking' in this sound.", + "choice_a": "[0.25, 10.0]", + "choice_b": "[2.67, 6.67]", + "choice_c": "[3.23, 8.2]", + "choice_d": "[3.62, 9.01]", + "answer_gt": "[0.25, 10.0]", + "answer_gt_word": "lots of dogs barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11125 + }, + { + "path": "Y1Mc63idyMZA.wav", + "question": "What are the comprehensive segments of 'an airplane or jet engine revving up' in time?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.85, 8.51]", + "choice_c": "[4.11, 8.8]", + "choice_d": "[8.26, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an airplane or jet engine revving up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11126 + }, + { + "path": "Y1Nl3caLibno.wav", + "question": "How many seconds does each full presence of 'engine running' cover?", + "choice_a": "[0.03, 9.08]", + "choice_b": "[0.0, 9.06]", + "choice_c": "[1.14, 2.88]", + "choice_d": "[2.67, 5.5]", + "answer_gt": "[0.03, 9.08]", + "answer_gt_word": "engine running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11127 + }, + { + "path": "Y1PMqP8xaMbk.wav", + "question": "Identify the periods when 'a crowd is murmuring' fully resonates in the sound.", + "choice_a": "[0.0, 8.13]", + "choice_b": "[0.0, 8.09]", + "choice_c": "[0.99, 5.5]", + "choice_d": "[5.13, 8.03]", + "answer_gt": "[0.0, 8.13]", + "answer_gt_word": "a crowd is murmuring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11128 + }, + { + "path": "Y1TIx22Pnt1Q.wav", + "question": "At which time markers does 'woman talking to a dog' fully unfold in the sound?", + "choice_a": "[0.58, 3.35]", + "choice_b": "[3.41, 3.78]", + "choice_c": "[6.26, 7.32]", + "choice_d": "[6.84, 9.66]", + "answer_gt": "[0.58, 3.35]", + "answer_gt_word": "woman talking to a dog", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11129 + }, + { + "path": "Y1X2Haq3e8wo.wav", + "question": "Identify the occurrences of 'an engine is idling' along with their time frames.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.08, 7.43]", + "choice_c": "[7.03, 7.61]", + "choice_d": "[7.34, 9.68]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11130 + }, + { + "path": "Y1_bLIXzVyio.wav", + "question": "What ranges in seconds capture 'a person whistling' in full?", + "choice_a": "[0.77, 2.11]", + "choice_b": "[0.0, 1.48]", + "choice_c": "[3.16, 4.18]", + "choice_d": "[3.16, 5.11]", + "answer_gt": "[0.77, 2.11]", + "answer_gt_word": "a person whistling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11131 + }, + { + "path": "Y1bE_vX8Oy80.wav", + "question": "At what time points is 'barking' fully present in the recording?", + "choice_a": "[0.0, 3.13]", + "choice_b": "[3.78, 10.0]", + "choice_c": "[0.01, 10.02]", + "choice_d": "[1.41, 6.24]", + "answer_gt": "[0.01, 10.02]", + "answer_gt_word": "barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11132 + }, + { + "path": "Y1jD2iUXynFo.wav", + "question": "What time segments showcase 'a vacuum is going' in its entirety?", + "choice_a": "[0.73, 10.0]", + "choice_b": "[1.73, 2.37]", + "choice_c": "[3.95, 4.86]", + "choice_d": "[6.4, 6.66]", + "answer_gt": "[0.73, 10.0]", + "answer_gt_word": "a vacuum is going", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11133 + }, + { + "path": "Y1slvoNgzBLE.wav", + "question": "List the specific seconds of 'an electronic signal' appearances.", + "choice_a": "[0.0, 2.59]", + "choice_b": "[5.22, 7.04]", + "choice_c": "[0.0, 2.59]", + "choice_d": "[3.39, 8.04]", + "answer_gt": "[0.0, 2.59]", + "answer_gt_word": "an electronic signal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11134 + }, + { + "path": "Y1tujDfcQZEs.wav", + "question": "What are the boundaries in seconds for each 'an electronic signal' presence?", + "choice_a": "[0.9, 5.76]", + "choice_b": "[0.0, 2.59]", + "choice_c": "[2.55, 3.78]", + "choice_d": "[7.05, 8.97]", + "answer_gt": "[0.0, 2.59]", + "answer_gt_word": "an electronic signal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11135 + }, + { + "path": "Y1vZNNVA7AMU.wav", + "question": "What are the exact times when 'a woman talks' is present in the clip?", + "choice_a": "[0.01, 9.19]", + "choice_b": "[6.15, 9.19]", + "choice_c": "[0.67, 1.72]", + "choice_d": "[2.97, 5.82]", + "answer_gt": "[0.01, 9.19]", + "answer_gt_word": "a woman talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11136 + }, + { + "path": "Y1vmlc5ba-OQ.wav", + "question": "Identify the complete instances of 'a woman talks' within the sound timeline.", + "choice_a": "[0.0, 3.92]", + "choice_b": "[4.45, 7.33]", + "choice_c": "[7.94, 9.93]", + "choice_d": "[0.01, 9.19]", + "answer_gt": "[0.01, 9.19]", + "answer_gt_word": "a woman talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11137 + }, + { + "path": "Y1xTMKVBMU0Y.wav", + "question": "What time segments delineate the presence of 'an aircraft rumbles outside'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.65, 6.76]", + "choice_c": "[3.21, 8.15]", + "choice_d": "[5.98, 8.33]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an aircraft rumbles outside", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11138 + }, + { + "path": "Y1xwOwK2T55c.wav", + "question": "What time segments showcase 'loud motor' in its entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.22, 7.27]", + "choice_c": "[8.29, 9.01]", + "choice_d": "[7.85, 9.88]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud motor", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11139 + }, + { + "path": "Y1y10mtbYQP8.wav", + "question": "How are the complete segments of 'loud motor' timed in this audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.49, 4.28]", + "choice_c": "[6.67, 8.07]", + "choice_d": "[1.59, 2.24]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud motor", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11140 + }, + { + "path": "Y1z_0FAzrbsU.wav", + "question": "Provide the start and end times of 'loud motor' within this audio.", + "choice_a": "[0.23, 1.3]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.89, 3.55]", + "choice_d": "[5.25, 9.13]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud motor", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11141 + }, + { + "path": "Y2-BcDQ7XO3Y.wav", + "question": "Where do we hear 'police sirens' in full within this sound recording?", + "choice_a": "[0.95, 2.59]", + "choice_b": "[0.86, 2.76]", + "choice_c": "[3.51, 5.61]", + "choice_d": "[8.03, 8.3]", + "answer_gt": "[0.95, 2.59]", + "answer_gt_word": "police sirens", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11142 + }, + { + "path": "Y20RwAJ4dtSU.wav", + "question": "What are the full time segments when 'dogs howling' appears in this sound?", + "choice_a": "[0.0, 2.84]", + "choice_b": "[7.04, 10.0]", + "choice_c": "[2.67, 10.0]", + "choice_d": "[1.63, 2.62]", + "answer_gt": "[2.67, 10.0]", + "answer_gt_word": "dogs howling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11143 + }, + { + "path": "Y21WMz5hG0Zs.wav", + "question": "How long does 'a man speaks' last each time it appears in the audio?", + "choice_a": "[1.15, 2.65]", + "choice_b": "[0.1, 10.0]", + "choice_c": "[0.12, 10.0]", + "choice_d": "[4.03, 5.19]", + "answer_gt": "[1.15, 2.65]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11144 + }, + { + "path": "Y23mRMsHfek0.wav", + "question": "How many seconds does each full presence of 'the dog barking' cover?", + "choice_a": "[2.22, 3.21]", + "choice_b": "[7.41, 7.73]", + "choice_c": "[5.16, 6.09]", + "choice_d": "[8.49, 8.86]", + "answer_gt": "[7.41, 7.73]", + "answer_gt_word": "the dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11145 + }, + { + "path": "Y24q-Od40cuw.wav", + "question": "What times mark the complete cycles of 'the dog barking' in the audio?", + "choice_a": "[1.45, 2.32]", + "choice_b": "[7.41, 7.73]", + "choice_c": "[6.12, 6.64]", + "choice_d": "[2.28, 3.62]", + "answer_gt": "[7.41, 7.73]", + "answer_gt_word": "the dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11146 + }, + { + "path": "Y2FZRNzNXEts.wav", + "question": "What are the comprehensive segments of 'laughing' in time?", + "choice_a": "[2.13, 4.16]", + "choice_b": "[5.32, 6.42]", + "choice_c": "[8.89, 9.27]", + "choice_d": "[4.73, 5.17]", + "answer_gt": "[4.73, 5.17]", + "answer_gt_word": "laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11147 + }, + { + "path": "Y2My3EAryutA.wav", + "question": "List the periods when 'a car is skidding' is completely played in the sound.", + "choice_a": "[0.0, 10.01]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[0.85, 5.52]", + "choice_d": "[2.42, 8.38]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "a car is skidding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11148 + }, + { + "path": "Y2PLQX17xFSE.wav", + "question": "Document the full appearances of 'tires squeal' in terms of seconds.", + "choice_a": "[7.87, 9.07]", + "choice_b": "[1.18, 5.59]", + "choice_c": "[1.8, 6.17]", + "choice_d": "[5.37, 6.62]", + "answer_gt": "[7.87, 9.07]", + "answer_gt_word": "tires squeal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11149 + }, + { + "path": "Y2Q5NJPpcrds.wav", + "question": "Document the full appearances of 'tires squeal' in terms of seconds.", + "choice_a": "[0.03, 7.01]", + "choice_b": "[7.87, 9.07]", + "choice_c": "[0.03, 6.97]", + "choice_d": "[7.47, 9.22]", + "answer_gt": "[7.87, 9.07]", + "answer_gt_word": "tires squeal", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11150 + }, + { + "path": "Y2R9naYtL0tM.wav", + "question": "What second marks correspond to the entire span of 'an engine stops'?", + "choice_a": "[0.0, 1.69]", + "choice_b": "[1.74, 3.65]", + "choice_c": "[4.74, 6.21]", + "choice_d": "[6.68, 9.02]", + "answer_gt": "[0.0, 1.69]", + "answer_gt_word": "an engine stops", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11151 + }, + { + "path": "Y2RTj-NQtrl0.wav", + "question": "Define the time intervals that fully contain 'a crash' in this sound.", + "choice_a": "[0.95, 1.4]", + "choice_b": "[8.64, 9.64]", + "choice_c": "[1.35, 3.37]", + "choice_d": "[5.14, 6.27]", + "answer_gt": "[0.95, 1.4]", + "answer_gt_word": "a crash", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11152 + }, + { + "path": "Y2SfiNxoZzQY.wav", + "question": "At what time intervals does 'a man speaks' occur in the sound file?", + "choice_a": "[7.79, 8.96]", + "choice_b": "[2.86, 3.26]", + "choice_c": "[4.1, 4.63]", + "choice_d": "[6.47, 7.16]", + "answer_gt": "[7.79, 8.96]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11153 + }, + { + "path": "Y2WQDTzY-v3Q.wav", + "question": "Enumerate the second-based occurrences of 'clicking starts' in full.", + "choice_a": "[1.54, 2.29]", + "choice_b": "[5.23, 10.02]", + "choice_c": "[0.36, 1.97]", + "choice_d": "[1.6, 4.42]", + "answer_gt": "[1.54, 2.29]", + "answer_gt_word": "clicking starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11154 + }, + { + "path": "Y2Zbl9J4vWns.wav", + "question": "Capture the timeline where 'clicking starts' is fully included in the sound.", + "choice_a": "[1.54, 2.29]", + "choice_b": "[2.88, 7.71]", + "choice_c": "[8.15, 10.0]", + "choice_d": "[0.22, 2.32]", + "answer_gt": "[1.54, 2.29]", + "answer_gt_word": "clicking starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11155 + }, + { + "path": "Y2b43Rkr-TO0.wav", + "question": "How long does 'clicking starts' last each time it appears in the audio?", + "choice_a": "[1.54, 2.29]", + "choice_b": "[5.26, 8.02]", + "choice_c": "[5.62, 7.55]", + "choice_d": "[8.62, 10.0]", + "answer_gt": "[1.54, 2.29]", + "answer_gt_word": "clicking starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11156 + }, + { + "path": "Y2c3gzj2jIfk.wav", + "question": "Identify the periods when 'clicking starts' fully resonates in the sound.", + "choice_a": "[0.0, 2.99]", + "choice_b": "[3.73, 4.34]", + "choice_c": "[1.54, 2.29]", + "choice_d": "[8.2, 10.0]", + "answer_gt": "[1.54, 2.29]", + "answer_gt_word": "clicking starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11157 + }, + { + "path": "Y2lP3y7kW9L0.wav", + "question": "What is the range in seconds for each full 'clicking starts' in the recording?", + "choice_a": "[0.0, 2.58]", + "choice_b": "[4.48, 4.77]", + "choice_c": "[6.03, 7.09]", + "choice_d": "[1.54, 2.29]", + "answer_gt": "[1.54, 2.29]", + "answer_gt_word": "clicking starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11158 + }, + { + "path": "Y2o93RmaqiLw.wav", + "question": "List the specific seconds of 'a train sounds its horn' appearances.", + "choice_a": "[3.38, 4.43]", + "choice_b": "[1.59, 3.79]", + "choice_c": "[3.24, 4.82]", + "choice_d": "[3.56, 7.59]", + "answer_gt": "[3.38, 4.43]", + "answer_gt_word": "a train sounds its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11159 + }, + { + "path": "Y2oOrBpVGJzc.wav", + "question": "What second marks correspond to the entire span of 'a train sounds its horn'?", + "choice_a": "[0.0, 0.38]", + "choice_b": "[3.38, 4.43]", + "choice_c": "[6.67, 8.2]", + "choice_d": "[9.67, 10.0]", + "answer_gt": "[3.38, 4.43]", + "answer_gt_word": "a train sounds its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11160 + }, + { + "path": "Y2r3RbvqKYEQ.wav", + "question": "At which seconds does 'a train sounds its horn' fully register in the sound?", + "choice_a": "[0.24, 4.55]", + "choice_b": "[3.38, 4.43]", + "choice_c": "[0.24, 4.55]", + "choice_d": "[6.0, 8.82]", + "answer_gt": "[3.38, 4.43]", + "answer_gt_word": "a train sounds its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11161 + }, + { + "path": "Y2sadF79hT7I.wav", + "question": "Identify the periods when 'a horn honks' fully resonates in the sound.", + "choice_a": "[0.02, 3.66]", + "choice_b": "[6.26, 10.0]", + "choice_c": "[3.64, 6.56]", + "choice_d": "[3.26, 4.5]", + "answer_gt": "[3.64, 6.56]", + "answer_gt_word": "a horn honks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11162 + }, + { + "path": "Y2uCe5wv_iQQ.wav", + "question": "At which seconds does 'a horn honks' fully register in the sound?", + "choice_a": "[0.0, 0.29]", + "choice_b": "[5.97, 7.16]", + "choice_c": "[3.64, 6.56]", + "choice_d": "[8.37, 8.83]", + "answer_gt": "[3.64, 6.56]", + "answer_gt_word": "a horn honks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11163 + }, + { + "path": "Y2wKP1A8On1E.wav", + "question": "At what time points is 'people speak' fully present in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.86, 7.13]", + "choice_d": "[5.39, 9.63]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people speak", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11164 + }, + { + "path": "Y2zGT7SeSHXo.wav", + "question": "Which sections of the sound contain 'a horse neighs loudly' in their entirety?", + "choice_a": "[0.01, 2.59]", + "choice_b": "[1.04, 5.05]", + "choice_c": "[8.08, 9.0]", + "choice_d": "[1.83, 7.84]", + "answer_gt": "[0.01, 2.59]", + "answer_gt_word": "a horse neighs loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11165 + }, + { + "path": "Y350OCezayrk.wav", + "question": "In what time spans is 'an engine starting' completely audible in this recording?", + "choice_a": "[0.0, 4.88]", + "choice_b": "[0.29, 6.27]", + "choice_c": "[0.45, 7.58]", + "choice_d": "[2.2, 7.79]", + "answer_gt": "[0.0, 4.88]", + "answer_gt_word": "an engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11166 + }, + { + "path": "Y38rpGMJA_HI.wav", + "question": "What are the precise second intervals of 'a fart' in this clip?", + "choice_a": "[1.16, 1.71]", + "choice_b": "[2.23, 3.17]", + "choice_c": "[3.27, 3.44]", + "choice_d": "[3.99, 4.72]", + "answer_gt": "[3.27, 3.44]", + "answer_gt_word": "a fart", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11167 + }, + { + "path": "Y3Bk--DQpmUs.wav", + "question": "What time segments delineate the presence of 'an adult male speaks'?", + "choice_a": "[0.0, 0.57]", + "choice_b": "[6.06, 9.99]", + "choice_c": "[0.58, 5.79]", + "choice_d": "[2.65, 9.51]", + "answer_gt": "[0.0, 0.57]", + "answer_gt_word": "an adult male speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11168 + }, + { + "path": "Y3Bl7Ko2xtmE.wav", + "question": "At which time markers does 'an engine is idling' fully unfold in the sound?", + "choice_a": "[0.0, 9.14]", + "choice_b": "[0.54, 6.62]", + "choice_c": "[6.43, 8.34]", + "choice_d": "[6.52, 8.56]", + "answer_gt": "[0.0, 9.14]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11169 + }, + { + "path": "Y3G2GK69lD4U.wav", + "question": "In what time spans is 'people chatter' completely audible in this recording?", + "choice_a": "[0.96, 4.2]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.19, 6.9]", + "choice_d": "[3.6, 8.32]", + "answer_gt": "[0.96, 4.2]", + "answer_gt_word": "people chatter", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11170 + }, + { + "path": "Y3Glt9M8l_OA.wav", + "question": "What are the full time segments when 'person laughing in background' appears in this sound?", + "choice_a": "[0.49, 3.09]", + "choice_b": "[6.53, 7.93]", + "choice_c": "[8.49, 9.16]", + "choice_d": "[2.63, 5.55]", + "answer_gt": "[8.49, 9.16]", + "answer_gt_word": "person laughing in background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11171 + }, + { + "path": "Y3IPOYt95c_Y.wav", + "question": "What times mark the complete cycles of 'a man speaks' in the audio?", + "choice_a": "[1.47, 2.77]", + "choice_b": "[4.25, 4.77]", + "choice_c": "[5.5, 6.04]", + "choice_d": "[6.68, 7.23]", + "answer_gt": "[1.47, 2.77]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11172 + }, + { + "path": "Y3KF0xSdYWeE.wav", + "question": "Provide the start and end times of 'people are talking' within this audio.", + "choice_a": "[0.58, 3.12]", + "choice_b": "[6.16, 7.21]", + "choice_c": "[8.11, 8.98]", + "choice_d": "[0.48, 10.0]", + "answer_gt": "[0.48, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11173 + }, + { + "path": "Y3KdlTPIqOpk.wav", + "question": "Which sections of the sound contain 'a person is speaking' in their entirety?", + "choice_a": "[0.0, 9.28]", + "choice_b": "[1.62, 2.37]", + "choice_c": "[0.22, 6.43]", + "choice_d": "[3.96, 7.83]", + "answer_gt": "[0.0, 9.28]", + "answer_gt_word": "a person is speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11174 + }, + { + "path": "Y3NNCGNtAlVE.wav", + "question": "List the specific seconds of 'an adult male speaks' appearances.", + "choice_a": "[0.03, 0.8]", + "choice_b": "[2.48, 10.0]", + "choice_c": "[1.41, 2.54]", + "choice_d": "[2.06, 6.28]", + "answer_gt": "[0.03, 0.8]", + "answer_gt_word": "an adult male speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11175 + }, + { + "path": "Y3OFi-JWcGK8.wav", + "question": "Define the time intervals that fully contain 'dog barking' in this sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[6.88, 10.02]", + "choice_d": "[1.73, 8.35]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11176 + }, + { + "path": "Y3OeT2w-6B94.wav", + "question": "Where is 'loud roar of vehicles passing by' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 9.82]", + "choice_b": "[8.22, 9.07]", + "choice_c": "[3.66, 8.5]", + "choice_d": "[7.8, 9.54]", + "answer_gt": "[0.0, 9.82]", + "answer_gt_word": "loud roar of vehicles passing by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11177 + }, + { + "path": "Y3P1EgAdpL98.wav", + "question": "Which sections of the sound contain 'loud roar of vehicles passing by' in their entirety?", + "choice_a": "[0.0, 0.62]", + "choice_b": "[1.79, 3.58]", + "choice_c": "[4.17, 4.48]", + "choice_d": "[0.0, 9.82]", + "answer_gt": "[0.0, 9.82]", + "answer_gt_word": "loud roar of vehicles passing by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11178 + }, + { + "path": "Y3PVSh7dth8Q.wav", + "question": "What is the range in seconds for each full 'wind against the microphone' in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.74, 5.95]", + "choice_c": "[4.55, 7.22]", + "choice_d": "[4.87, 8.7]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind against the microphone", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11179 + }, + { + "path": "Y3SFet_JSgfY.wav", + "question": "Where is 'a car is revving up its engine' consistently heard throughout the sound?", + "choice_a": "[0.02, 6.05]", + "choice_b": "[2.13, 7.59]", + "choice_c": "[3.9, 9.41]", + "choice_d": "[6.06, 9.88]", + "answer_gt": "[0.02, 6.05]", + "answer_gt_word": "a car is revving up its engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11180 + }, + { + "path": "Y3TZTVzCsPrc.wav", + "question": "At what time intervals does 'a whirring engine suddenly' occur in the sound file?", + "choice_a": "[0.0, 7.3]", + "choice_b": "[8.62, 9.45]", + "choice_c": "[7.23, 8.55]", + "choice_d": "[0.35, 7.09]", + "answer_gt": "[0.0, 7.3]", + "answer_gt_word": "a whirring engine suddenly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11181 + }, + { + "path": "Y3VfzEumJgcQ.wav", + "question": "List the specific seconds of 'a whirring engine suddenly' appearances.", + "choice_a": "[0.0, 0.71]", + "choice_b": "[1.58, 2.57]", + "choice_c": "[0.0, 7.3]", + "choice_d": "[6.63, 7.96]", + "answer_gt": "[0.0, 7.3]", + "answer_gt_word": "a whirring engine suddenly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11182 + }, + { + "path": "Y3ZAtv4S5o5Q.wav", + "question": "Where do we hear 'a whirring engine suddenly' in full within this sound recording?", + "choice_a": "[1.29, 5.29]", + "choice_b": "[5.54, 6.81]", + "choice_c": "[7.18, 8.6]", + "choice_d": "[0.0, 7.3]", + "answer_gt": "[0.0, 7.3]", + "answer_gt_word": "a whirring engine suddenly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11183 + }, + { + "path": "Y3fQyxJQXF3g.wav", + "question": "What are the comprehensive timeframes of 'a man speaks' in this clip?", + "choice_a": "[0.0, 7.05]", + "choice_b": "[9.23, 9.81]", + "choice_c": "[7.05, 7.81]", + "choice_d": "[7.74, 8.81]", + "answer_gt": "[7.05, 7.81]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11184 + }, + { + "path": "Y3kNCkzSBrfM.wav", + "question": "What ranges in seconds capture 'motors rev' in full?", + "choice_a": "[0.0, 3.2]", + "choice_b": "[3.81, 6.34]", + "choice_c": "[8.54, 10.0]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "motors rev", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11185 + }, + { + "path": "Y3leh-_Xki-0.wav", + "question": "What are the exact times when 'a dog barks' is present in the clip?", + "choice_a": "[0.0, 0.48]", + "choice_b": "[4.07, 5.64]", + "choice_c": "[8.83, 9.05]", + "choice_d": "[3.43, 5.5]", + "answer_gt": "[3.43, 5.5]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11186 + }, + { + "path": "Y3pLkwWZxvYo.wav", + "question": "Capture the timeline where 'a dog barks' is fully included in the sound.", + "choice_a": "[0.46, 1.89]", + "choice_b": "[6.58, 6.85]", + "choice_c": "[3.43, 5.5]", + "choice_d": "[2.69, 3.61]", + "answer_gt": "[3.43, 5.5]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11187 + }, + { + "path": "Y3rfkzWJqq7U.wav", + "question": "What are the full time segments when 'an engine is started up' appears in this sound?", + "choice_a": "[1.78, 4.94]", + "choice_b": "[0.02, 2.42]", + "choice_c": "[6.74, 8.25]", + "choice_d": "[5.26, 8.99]", + "answer_gt": "[1.78, 4.94]", + "answer_gt_word": "an engine is started up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11188 + }, + { + "path": "Y3ujCBi8VbLc.wav", + "question": "What time segments showcase 'an engine is started up' in its entirety?", + "choice_a": "[1.78, 4.94]", + "choice_b": "[6.72, 7.15]", + "choice_c": "[2.74, 3.4]", + "choice_d": "[4.97, 5.64]", + "answer_gt": "[1.78, 4.94]", + "answer_gt_word": "an engine is started up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11189 + }, + { + "path": "Y3wabvpoDwKU.wav", + "question": "In what time spans is 'a woman giggling' completely audible in this recording?", + "choice_a": "[0.73, 1.71]", + "choice_b": "[3.02, 3.94]", + "choice_c": "[4.58, 5.19]", + "choice_d": "[6.0, 6.56]", + "answer_gt": "[6.0, 6.56]", + "answer_gt_word": "a woman giggling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11190 + }, + { + "path": "Y40KPvO4y9M8.wav", + "question": "In what time spans is 'dogs are barking' completely audible in this recording?", + "choice_a": "[1.87, 6.2]", + "choice_b": "[0.0, 0.29]", + "choice_c": "[3.95, 5.07]", + "choice_d": "[6.39, 8.12]", + "answer_gt": "[1.87, 6.2]", + "answer_gt_word": "dogs are barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11191 + }, + { + "path": "Y45KEHJraS_I.wav", + "question": "How is 'dogs are barking' distributed over the sound's duration?", + "choice_a": "[1.11, 3.17]", + "choice_b": "[1.87, 6.2]", + "choice_c": "[8.2, 9.54]", + "choice_d": "[2.81, 3.75]", + "answer_gt": "[1.87, 6.2]", + "answer_gt_word": "dogs are barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11192 + }, + { + "path": "Y47tADKZxUnM.wav", + "question": "How is 'the wind is blowing' distributed over the sound's duration?", + "choice_a": "[0.35, 2.74]", + "choice_b": "[3.52, 5.46]", + "choice_c": "[6.44, 9.38]", + "choice_d": "[5.44, 10.0]", + "answer_gt": "[5.44, 10.0]", + "answer_gt_word": "the wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11193 + }, + { + "path": "Y487WZLlab7s.wav", + "question": "Where is 'the wind is blowing' found in its entirety within the audio's timeline?", + "choice_a": "[0.12, 1.18]", + "choice_b": "[5.44, 10.0]", + "choice_c": "[7.69, 9.14]", + "choice_d": "[5.37, 5.8]", + "answer_gt": "[5.44, 10.0]", + "answer_gt_word": "the wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11194 + }, + { + "path": "Y4AvF_quFNSg.wav", + "question": "What are the precise second intervals of 'vroom of an engine accelerating' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.1, 4.08]", + "choice_c": "[2.15, 6.44]", + "choice_d": "[3.3, 9.36]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "vroom of an engine accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11195 + }, + { + "path": "Y4CZ6vWHPVr0.wav", + "question": "At what time points is 'revs up' fully present in the recording?", + "choice_a": "[0.01, 2.86]", + "choice_b": "[4.28, 10.0]", + "choice_c": "[2.78, 4.3]", + "choice_d": "[0.11, 4.32]", + "answer_gt": "[2.78, 4.3]", + "answer_gt_word": "revs up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11196 + }, + { + "path": "Y4LV2zS3hJ_I.wav", + "question": "What are the precise second intervals of 'a car is idling' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.32, 1.92]", + "choice_c": "[3.48, 4.98]", + "choice_d": "[5.64, 7.37]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a car is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11197 + }, + { + "path": "Y4N01YFev3lo.wav", + "question": "How is 'a car is idling' distributed over the sound's duration?", + "choice_a": "[0.0, 1.28]", + "choice_b": "[2.31, 2.92]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[7.31, 8.24]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a car is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11198 + }, + { + "path": "Y4QEON5KWC1o.wav", + "question": "What time segments showcase 'cracking of leaves' in its entirety?", + "choice_a": "[0.01, 4.2]", + "choice_b": "[8.54, 8.85]", + "choice_c": "[9.72, 10.0]", + "choice_d": "[1.24, 1.94]", + "answer_gt": "[0.01, 4.2]", + "answer_gt_word": "cracking of leaves", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11199 + }, + { + "path": "Y4RU8_lp0Z2E.wav", + "question": "What are the comprehensive segments of 'talking' in time?", + "choice_a": "[0.41, 1.52]", + "choice_b": "[4.0, 4.33]", + "choice_c": "[7.38, 9.12]", + "choice_d": "[8.25, 9.11]", + "answer_gt": "[7.38, 9.12]", + "answer_gt_word": "talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11200 + }, + { + "path": "Y4UTdag5Lp0o.wav", + "question": "What are the boundaries in seconds for each 'an electric saw cutting wood' presence?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.87, 5.58]", + "choice_c": "[1.24, 6.24]", + "choice_d": "[4.78, 9.26]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an electric saw cutting wood", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11201 + }, + { + "path": "Y4YUp45FEi5U.wav", + "question": "What timestamps mark the full occurrences of 'an electric saw cutting wood'?", + "choice_a": "[0.0, 4.9]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.29, 0.62]", + "choice_d": "[4.99, 5.44]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an electric saw cutting wood", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11202 + }, + { + "path": "Y4ZdX1Hl6mhs.wav", + "question": "What are the intervals in seconds of each 'an electric saw cutting wood' appearance?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.79, 4.13]", + "choice_c": "[5.33, 6.09]", + "choice_d": "[6.29, 8.32]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an electric saw cutting wood", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11203 + }, + { + "path": "Y4dgvqmpMORQ.wav", + "question": "Locate every segment where 'people are talking' is completely audible.", + "choice_a": "[0.86, 2.53]", + "choice_b": "[3.35, 4.42]", + "choice_c": "[9.16, 9.53]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11204 + }, + { + "path": "Y4iwHOSTWd1s.wav", + "question": "Where is 'an adult female speaks in the foreground for a moment' consistently heard throughout the sound?", + "choice_a": "[0.0, 4.73]", + "choice_b": "[6.91, 8.55]", + "choice_c": "[0.95, 4.68]", + "choice_d": "[4.58, 10.0]", + "answer_gt": "[0.95, 4.68]", + "answer_gt_word": "an adult female speaks in the foreground for a moment", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11205 + }, + { + "path": "Y4v_R71cDy5Q.wav", + "question": "How is 'some splashing' distributed over the sound's duration?", + "choice_a": "[0.0, 0.43]", + "choice_b": "[1.0, 1.41]", + "choice_c": "[0.0, 9.7]", + "choice_d": "[3.29, 6.83]", + "answer_gt": "[0.0, 9.7]", + "answer_gt_word": "some splashing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11206 + }, + { + "path": "Y4w16MdaZQpU.wav", + "question": "Where is 'high pitched electrical vibrations' consistently heard throughout the sound?", + "choice_a": "[0.0, 4.04]", + "choice_b": "[6.87, 7.27]", + "choice_c": "[0.54, 2.08]", + "choice_d": "[2.97, 4.03]", + "answer_gt": "[0.0, 4.04]", + "answer_gt_word": "high pitched electrical vibrations", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11207 + }, + { + "path": "Y4xgFvYLnovM.wav", + "question": "At what seconds do the full iterations of 'high pitched electrical vibrations' start and end?", + "choice_a": "[1.78, 2.82]", + "choice_b": "[4.94, 5.96]", + "choice_c": "[4.93, 5.17]", + "choice_d": "[0.0, 4.04]", + "answer_gt": "[0.0, 4.04]", + "answer_gt_word": "high pitched electrical vibrations", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11208 + }, + { + "path": "Y5-4pJcg-N6s.wav", + "question": "Capture the timeline where 'high pitched electrical vibrations' is fully included in the sound.", + "choice_a": "[0.0, 1.22]", + "choice_b": "[6.44, 6.91]", + "choice_c": "[0.0, 4.04]", + "choice_d": "[1.34, 5.0]", + "answer_gt": "[0.0, 4.04]", + "answer_gt_word": "high pitched electrical vibrations", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11209 + }, + { + "path": "Y5-PJ49JuuRQ.wav", + "question": "How many seconds does each full presence of 'loud engine revving' cover?", + "choice_a": "[0.0, 9.07]", + "choice_b": "[0.28, 6.04]", + "choice_c": "[1.92, 8.01]", + "choice_d": "[2.66, 8.21]", + "answer_gt": "[0.0, 9.07]", + "answer_gt_word": "loud engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11210 + }, + { + "path": "Y513JkjMF_Os.wav", + "question": "What are the distinct time ranges of 'loud engine revving' within the audio?", + "choice_a": "[1.47, 2.38]", + "choice_b": "[0.0, 9.07]", + "choice_c": "[0.0, 1.21]", + "choice_d": "[2.09, 3.81]", + "answer_gt": "[0.0, 9.07]", + "answer_gt_word": "loud engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11211 + }, + { + "path": "Y5CVI9p6nNuQ.wav", + "question": "What ranges in seconds capture 'a cat meowing' in full?", + "choice_a": "[3.93, 4.62]", + "choice_b": "[6.05, 6.88]", + "choice_c": "[8.21, 9.27]", + "choice_d": "[5.51, 5.8]", + "answer_gt": "[3.93, 4.62]", + "answer_gt_word": "a cat meowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11212 + }, + { + "path": "Y5HdTmxH9A-k.wav", + "question": "At what seconds do the full iterations of 'a cat meowing' start and end?", + "choice_a": "[0.01, 1.43]", + "choice_b": "[2.15, 3.08]", + "choice_c": "[4.61, 5.96]", + "choice_d": "[3.93, 4.62]", + "answer_gt": "[3.93, 4.62]", + "answer_gt_word": "a cat meowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11213 + }, + { + "path": "Y5KmU6t2Csi0.wav", + "question": "What ranges in seconds capture 'it idles' in full?", + "choice_a": "[0.01, 3.42]", + "choice_b": "[7.71, 10.0]", + "choice_c": "[3.14, 7.51]", + "choice_d": "[2.49, 3.0]", + "answer_gt": "[3.14, 7.51]", + "answer_gt_word": "it idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11214 + }, + { + "path": "Y5MHiwsp7Dd0.wav", + "question": "What are the specific start and stop times of 'it idles' in the sound?", + "choice_a": "[0.0, 3.17]", + "choice_b": "[6.66, 10.0]", + "choice_c": "[3.14, 7.51]", + "choice_d": "[3.25, 6.1]", + "answer_gt": "[3.14, 7.51]", + "answer_gt_word": "it idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11215 + }, + { + "path": "Y5SFxo4VFhBM.wav", + "question": "What are the full time segments when 'it idles' appears in this sound?", + "choice_a": "[3.14, 7.51]", + "choice_b": "[1.62, 2.14]", + "choice_c": "[2.88, 3.2]", + "choice_d": "[3.57, 4.83]", + "answer_gt": "[3.14, 7.51]", + "answer_gt_word": "it idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11216 + }, + { + "path": "Y5T6asFC8WDI.wav", + "question": "What timestamps mark the full occurrences of 'man laughing'?", + "choice_a": "[1.35, 1.77]", + "choice_b": "[2.52, 2.93]", + "choice_c": "[3.95, 4.86]", + "choice_d": "[0.28, 1.99]", + "answer_gt": "[0.28, 1.99]", + "answer_gt_word": "man laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11217 + }, + { + "path": "Y5cWOyow_Drs.wav", + "question": "What time segments delineate the presence of 'man laughing'?", + "choice_a": "[3.48, 4.15]", + "choice_b": "[5.08, 5.6]", + "choice_c": "[0.28, 1.99]", + "choice_d": "[8.18, 9.25]", + "answer_gt": "[0.28, 1.99]", + "answer_gt_word": "man laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11218 + }, + { + "path": "Y5jpR3i8moAE.wav", + "question": "What are the boundaries in seconds for each 'water moving' presence?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[1.61, 6.55]", + "choice_d": "[3.39, 8.6]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "water moving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11219 + }, + { + "path": "Y5kiqvp63bTU.wav", + "question": "Where does 'an engine is idling' appear from beginning to end in the sound?", + "choice_a": "[7.89, 10.0]", + "choice_b": "[0.0, 4.2]", + "choice_c": "[0.72, 5.21]", + "choice_d": "[1.84, 5.72]", + "answer_gt": "[7.89, 10.0]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11220 + }, + { + "path": "Y5tOUOfmpL0c.wav", + "question": "Identify the complete instances of 'banging noise in background' within the sound timeline.", + "choice_a": "[0.08, 10.0]", + "choice_b": "[0.07, 10.0]", + "choice_c": "[0.11, 8.68]", + "choice_d": "[7.55, 9.47]", + "answer_gt": "[0.08, 10.0]", + "answer_gt_word": "banging noise in background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11221 + }, + { + "path": "Y5wNBcUVUbsk.wav", + "question": "At what timestamps does the complete 'wind blowing hard' occur?", + "choice_a": "[0.01, 2.88]", + "choice_b": "[2.9, 10.0]", + "choice_c": "[5.44, 7.26]", + "choice_d": "[7.1, 8.64]", + "answer_gt": "[0.01, 2.88]", + "answer_gt_word": "wind blowing hard", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11222 + }, + { + "path": "Y5wfJoXRDF2g.wav", + "question": "What time segments delineate the presence of 'wind blowing hard'?", + "choice_a": "[1.1, 2.54]", + "choice_b": "[4.18, 5.9]", + "choice_c": "[0.01, 2.88]", + "choice_d": "[5.97, 7.31]", + "answer_gt": "[0.01, 2.88]", + "answer_gt_word": "wind blowing hard", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11223 + }, + { + "path": "Y5xOlEfsD0_8.wav", + "question": "Provide the start and end times of 'a car engine starts' within this audio.", + "choice_a": "[1.15, 2.15]", + "choice_b": "[1.74, 9.07]", + "choice_c": "[4.07, 8.02]", + "choice_d": "[7.21, 8.04]", + "answer_gt": "[1.15, 2.15]", + "answer_gt_word": "a car engine starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11224 + }, + { + "path": "Y5zPgffCgLk0.wav", + "question": "What second marks correspond to the entire span of 'a car engine starts'?", + "choice_a": "[0.01, 0.62]", + "choice_b": "[1.15, 2.15]", + "choice_c": "[7.76, 8.43]", + "choice_d": "[5.94, 6.3]", + "answer_gt": "[1.15, 2.15]", + "answer_gt_word": "a car engine starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11225 + }, + { + "path": "Y5zcCl0Lxt24.wav", + "question": "List the specific seconds of 'a car engine starts' appearances.", + "choice_a": "[0.0, 1.16]", + "choice_b": "[1.67, 2.5]", + "choice_c": "[2.99, 5.46]", + "choice_d": "[1.15, 2.15]", + "answer_gt": "[1.15, 2.15]", + "answer_gt_word": "a car engine starts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11226 + }, + { + "path": "Y6-UwJ4gAjO8.wav", + "question": "What are the precise second intervals of 'a alarm sounding' in this clip?", + "choice_a": "[0.0, 8.66]", + "choice_b": "[1.91, 6.14]", + "choice_c": "[2.72, 7.06]", + "choice_d": "[3.54, 9.21]", + "answer_gt": "[0.0, 8.66]", + "answer_gt_word": "a alarm sounding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11227 + }, + { + "path": "Y62nf3wIQrS8.wav", + "question": "What second marks correspond to the entire span of 'idling engine sound'?", + "choice_a": "[0.01, 7.1]", + "choice_b": "[8.16, 10.0]", + "choice_c": "[6.86, 10.0]", + "choice_d": "[3.01, 4.96]", + "answer_gt": "[0.01, 7.1]", + "answer_gt_word": "idling engine sound", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11228 + }, + { + "path": "Y64EcdCriKFU.wav", + "question": "How long in seconds does 'small dogs bark' last from start to end?", + "choice_a": "[0.01, 7.7]", + "choice_b": "[0.01, 7.66]", + "choice_c": "[2.25, 7.13]", + "choice_d": "[4.14, 9.76]", + "answer_gt": "[0.01, 7.7]", + "answer_gt_word": "small dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11229 + }, + { + "path": "Y6G11Ve9d1MQ.wav", + "question": "How long in seconds does 'an adult male speaks in the background' last from start to end?", + "choice_a": "[0.76, 9.87]", + "choice_b": "[3.26, 10.01]", + "choice_c": "[0.24, 5.11]", + "choice_d": "[3.76, 8.51]", + "answer_gt": "[0.76, 9.87]", + "answer_gt_word": "an adult male speaks in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11230 + }, + { + "path": "Y6K4wRfU5JEQ.wav", + "question": "In what segments can 'wind noise' be heard from start to finish?", + "choice_a": "[0.0, 7.83]", + "choice_b": "[9.78, 10.0]", + "choice_c": "[7.65, 10.0]", + "choice_d": "[1.11, 3.12]", + "answer_gt": "[7.65, 10.0]", + "answer_gt_word": "wind noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11231 + }, + { + "path": "Y6SxVYRx-g80.wav", + "question": "How is 'several dogs bark and whimper' distributed over the sound's duration?", + "choice_a": "[0.5, 8.84]", + "choice_b": "[7.59, 9.79]", + "choice_c": "[1.14, 8.16]", + "choice_d": "[6.06, 9.02]", + "answer_gt": "[0.5, 8.84]", + "answer_gt_word": "several dogs bark and whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11232 + }, + { + "path": "Y6TP8w9CcdHk.wav", + "question": "What are the comprehensive timeframes of 'fast , running water' in this clip?", + "choice_a": "[0.0, 9.97]", + "choice_b": "[1.23, 9.38]", + "choice_c": "[1.23, 9.37]", + "choice_d": "[0.2, 7.78]", + "answer_gt": "[0.0, 9.97]", + "answer_gt_word": "fast , running water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11233 + }, + { + "path": "Y6VJWBMh416c.wav", + "question": "Enumerate the second-based occurrences of 'a dog barks' in full.", + "choice_a": "[0.02, 3.1]", + "choice_b": "[3.26, 4.44]", + "choice_c": "[6.79, 7.7]", + "choice_d": "[4.56, 4.92]", + "answer_gt": "[0.02, 3.1]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11234 + }, + { + "path": "Y6WteI_2ipko.wav", + "question": "What times mark the complete cycles of 'a young female laughs' in the audio?", + "choice_a": "[6.96, 8.0]", + "choice_b": "[1.6, 3.06]", + "choice_c": "[5.42, 6.25]", + "choice_d": "[8.01, 9.46]", + "answer_gt": "[6.96, 8.0]", + "answer_gt_word": "a young female laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11235 + }, + { + "path": "Y6aiLRDp5hco.wav", + "question": "At what timestamps does the complete 'woman speaking whining dog' occur?", + "choice_a": "[0.0, 1.55]", + "choice_b": "[3.6, 6.17]", + "choice_c": "[7.16, 9.88]", + "choice_d": "[3.43, 4.43]", + "answer_gt": "[3.43, 4.43]", + "answer_gt_word": "woman speaking whining dog", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11236 + }, + { + "path": "Y6dLV_94FiNM.wav", + "question": "Identify the complete instances of 'people speaking' within the sound timeline.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.14, 8.79]", + "choice_c": "[6.85, 8.89]", + "choice_d": "[7.44, 9.65]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11237 + }, + { + "path": "Y6gisYWJffTI.wav", + "question": "Define the time intervals that fully contain 'an aircraft engine roars' in this sound.", + "choice_a": "[0.01, 10.02]", + "choice_b": "[0.31, 5.31]", + "choice_c": "[2.3, 7.97]", + "choice_d": "[5.14, 9.46]", + "answer_gt": "[0.01, 10.02]", + "answer_gt_word": "an aircraft engine roars", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11238 + }, + { + "path": "Y6hRncnuF1NY.wav", + "question": "At what timestamps does the complete 'alarm bells ringing' occur?", + "choice_a": "[0.0, 0.91]", + "choice_b": "[5.04, 6.04]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[3.47, 4.26]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "alarm bells ringing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11239 + }, + { + "path": "Y6i-BHYm4hYI.wav", + "question": "How are the complete segments of 'alarm bells ringing' timed in this audio?", + "choice_a": "[1.25, 1.54]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[1.9, 2.55]", + "choice_d": "[2.98, 4.23]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "alarm bells ringing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11240 + }, + { + "path": "Y6ipK0W9CbK0.wav", + "question": "Define the time intervals that fully contain 'people speaking and whimpering' in this sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.31, 8.36]", + "choice_c": "[3.0, 8.58]", + "choice_d": "[6.76, 9.42]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people speaking and whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11241 + }, + { + "path": "Y6iwOOiYQAt0.wav", + "question": "At which seconds does 'people speaking and whimpering' fully register in the sound?", + "choice_a": "[0.0, 1.29]", + "choice_b": "[5.69, 6.86]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[4.53, 5.27]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people speaking and whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11242 + }, + { + "path": "Y6jLu70WHfEU.wav", + "question": "Identify the periods when 'people laugh' fully resonates in the sound.", + "choice_a": "[5.26, 9.01]", + "choice_b": "[1.1, 1.79]", + "choice_c": "[2.64, 2.89]", + "choice_d": "[4.85, 5.07]", + "answer_gt": "[5.26, 9.01]", + "answer_gt_word": "people laugh", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11243 + }, + { + "path": "Y6jRpHlZgVy8.wav", + "question": "What are all the complete time segments of 'people laugh' in this sound?", + "choice_a": "[0.0, 0.7]", + "choice_b": "[1.1, 1.42]", + "choice_c": "[5.26, 9.01]", + "choice_d": "[9.39, 10.0]", + "answer_gt": "[5.26, 9.01]", + "answer_gt_word": "people laugh", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11244 + }, + { + "path": "Y6ohU5qMW9yo.wav", + "question": "How many seconds does 'people laugh' span in each occurrence?", + "choice_a": "[0.0, 0.38]", + "choice_b": "[1.18, 2.38]", + "choice_c": "[5.26, 9.01]", + "choice_d": "[0.39, 2.91]", + "answer_gt": "[5.26, 9.01]", + "answer_gt_word": "people laugh", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11245 + }, + { + "path": "Y6sXAXmSQ5To.wav", + "question": "Where is 'a loud car engine revving up' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 4.83]", + "choice_b": "[1.26, 1.63]", + "choice_c": "[4.8, 9.85]", + "choice_d": "[1.25, 1.62]", + "answer_gt": "[0.0, 4.83]", + "answer_gt_word": "a loud car engine revving up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11246 + }, + { + "path": "Y6t9R8V0HNXM.wav", + "question": "What are the distinct time ranges of 'a loud car engine revving up' within the audio?", + "choice_a": "[3.77, 4.41]", + "choice_b": "[7.53, 8.3]", + "choice_c": "[0.0, 4.83]", + "choice_d": "[4.66, 9.87]", + "answer_gt": "[0.0, 4.83]", + "answer_gt_word": "a loud car engine revving up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11247 + }, + { + "path": "Y6wtUZBZIjBs.wav", + "question": "At what points does 'a loud car engine revving up' exist completely in the timeline?", + "choice_a": "[0.09, 0.78]", + "choice_b": "[2.14, 2.42]", + "choice_c": "[0.0, 4.83]", + "choice_d": "[8.22, 8.54]", + "answer_gt": "[0.0, 4.83]", + "answer_gt_word": "a loud car engine revving up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11248 + }, + { + "path": "Y72JnuHYZuVE.wav", + "question": "Where does 'something is being snapped' appear from beginning to end in the sound?", + "choice_a": "[5.86, 9.08]", + "choice_b": "[0.55, 7.62]", + "choice_c": "[8.71, 9.45]", + "choice_d": "[0.55, 7.71]", + "answer_gt": "[5.86, 9.08]", + "answer_gt_word": "something is being snapped", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11249 + }, + { + "path": "Y73Vz2uj8iKY.wav", + "question": "Provide the start and end times of 'an engine is idling outside' within this audio.", + "choice_a": "[0.0, 4.13]", + "choice_b": "[6.09, 8.79]", + "choice_c": "[9.7, 10.0]", + "choice_d": "[1.12, 9.1]", + "answer_gt": "[0.0, 4.13]", + "answer_gt_word": "an engine is idling outside", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11250 + }, + { + "path": "Y73jR8wkpTt8.wav", + "question": "At what time points is 'water splashes' fully present in the recording?", + "choice_a": "[0.0, 8.51]", + "choice_b": "[8.52, 8.86]", + "choice_c": "[2.25, 5.64]", + "choice_d": "[2.92, 9.11]", + "answer_gt": "[0.0, 8.51]", + "answer_gt_word": "water splashes", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11251 + }, + { + "path": "Y7AScPytrpac.wav", + "question": "Where is 'several dogs bark' consistently heard throughout the sound?", + "choice_a": "[0.0, 9.61]", + "choice_b": "[1.5, 6.78]", + "choice_c": "[2.31, 7.41]", + "choice_d": "[5.36, 8.38]", + "answer_gt": "[0.0, 9.61]", + "answer_gt_word": "several dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11252 + }, + { + "path": "Y7Ay-gVQ4oSc.wav", + "question": "Where is 'a cat meows once' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 0.77]", + "choice_b": "[1.54, 2.87]", + "choice_c": "[3.42, 3.66]", + "choice_d": "[5.55, 6.27]", + "answer_gt": "[3.42, 3.66]", + "answer_gt_word": "a cat meows once", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11253 + }, + { + "path": "Y7Cm9bUZNVuk.wav", + "question": "Provide the start and end times of 'two chainsaws are being used' within this audio.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.66, 9.3]", + "choice_c": "[0.62, 7.88]", + "choice_d": "[5.08, 8.78]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "two chainsaws are being used", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11254 + }, + { + "path": "Y7EWWhDSFreY.wav", + "question": "How many seconds does each full presence of 'a female speaking' cover?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.21, 2.98]", + "choice_c": "[3.85, 10.02]", + "choice_d": "[5.11, 6.57]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a female speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11255 + }, + { + "path": "Y7GoQ8HaVAnA.wav", + "question": "How many seconds does 'several sirens are blaring very loudly' span in each occurrence?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.03, 4.39]", + "choice_c": "[1.39, 7.39]", + "choice_d": "[2.78, 8.69]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "several sirens are blaring very loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11256 + }, + { + "path": "Y7HyTIyWWRu8.wav", + "question": "How many seconds does each full presence of 'a dog growls' cover?", + "choice_a": "[0.43, 2.05]", + "choice_b": "[3.5, 4.18]", + "choice_c": "[5.71, 7.26]", + "choice_d": "[8.55, 10.0]", + "answer_gt": "[8.55, 10.0]", + "answer_gt_word": "a dog growls", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11257 + }, + { + "path": "Y7Ivl23eqQto.wav", + "question": "What times mark the complete cycles of 'barking' in the audio?", + "choice_a": "[0.0, 0.96]", + "choice_b": "[2.94, 3.37]", + "choice_c": "[3.41, 3.7]", + "choice_d": "[5.09, 5.74]", + "answer_gt": "[2.94, 3.37]", + "answer_gt_word": "barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11258 + }, + { + "path": "Y7X3i7llSK9U.wav", + "question": "What time segments showcase 'a train is blowing its horn' in its entirety?", + "choice_a": "[0.08, 5.83]", + "choice_b": "[0.01, 5.13]", + "choice_c": "[0.34, 6.58]", + "choice_d": "[3.01, 9.55]", + "answer_gt": "[0.08, 5.83]", + "answer_gt_word": "a train is blowing its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11259 + }, + { + "path": "Y7ZeCyUZT8Dg.wav", + "question": "What are the specific start and stop times of 'a large jet engine idling' in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.88, 7.27]", + "choice_c": "[5.94, 9.52]", + "choice_d": "[6.78, 9.84]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a large jet engine idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11260 + }, + { + "path": "Y7aifdCWotdI.wav", + "question": "In what segments can 'footsteps' be heard from start to finish?", + "choice_a": "[2.06, 8.41]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.41, 6.7]", + "choice_d": "[6.69, 7.82]", + "answer_gt": "[2.06, 8.41]", + "answer_gt_word": "footsteps", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11261 + }, + { + "path": "Y7cUkRQrNSIE.wav", + "question": "Where is 'cat meow' consistently heard throughout the sound?", + "choice_a": "[2.4, 3.57]", + "choice_b": "[4.58, 5.91]", + "choice_c": "[7.42, 8.44]", + "choice_d": "[8.85, 9.63]", + "answer_gt": "[8.85, 9.63]", + "answer_gt_word": "cat meow", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11262 + }, + { + "path": "Y7dNJWByoz10.wav", + "question": "What are the full time segments when 'a dog is panting heavily' appears in this sound?", + "choice_a": "[3.06, 10.0]", + "choice_b": "[5.35, 5.99]", + "choice_c": "[6.56, 7.34]", + "choice_d": "[8.23, 10.0]", + "answer_gt": "[3.06, 10.0]", + "answer_gt_word": "a dog is panting heavily", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11264 + }, + { + "path": "Y7h-wCrAdSNY.wav", + "question": "Identify the complete instances of 'a dog is panting heavily' within the sound timeline.", + "choice_a": "[0.0, 2.07]", + "choice_b": "[3.06, 7.18]", + "choice_c": "[3.06, 10.0]", + "choice_d": "[0.95, 1.86]", + "answer_gt": "[3.06, 10.0]", + "answer_gt_word": "a dog is panting heavily", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11265 + }, + { + "path": "Y7o5Q7fKYXL4.wav", + "question": "Where do we hear 'a man speaking' in full within this sound recording?", + "choice_a": "[1.84, 4.01]", + "choice_b": "[5.43, 6.74]", + "choice_c": "[2.67, 7.12]", + "choice_d": "[5.61, 9.96]", + "answer_gt": "[1.84, 4.01]", + "answer_gt_word": "a man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11266 + }, + { + "path": "Y7oqLwoYw-hQ.wav", + "question": "Provide the start and end times of 'a vehicle is revving up' within this audio.", + "choice_a": "[0.05, 10.0]", + "choice_b": "[2.68, 4.24]", + "choice_c": "[1.28, 5.46]", + "choice_d": "[5.09, 5.68]", + "answer_gt": "[0.05, 10.0]", + "answer_gt_word": "a vehicle is revving up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11267 + }, + { + "path": "Y7ynQ9wd6neQ.wav", + "question": "What times mark the complete cycles of 'a car engine is idling deeply' in the audio?", + "choice_a": "[0.0, 6.93]", + "choice_b": "[6.93, 10.0]", + "choice_c": "[0.48, 7.71]", + "choice_d": "[5.1, 8.21]", + "answer_gt": "[0.0, 6.93]", + "answer_gt_word": "a car engine is idling deeply", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11268 + }, + { + "path": "Y81C5ysSrqyY.wav", + "question": "What time segments showcase 'a car engine is idling deeply' in its entirety?", + "choice_a": "[0.0, 1.09]", + "choice_b": "[3.65, 4.55]", + "choice_c": "[0.0, 6.93]", + "choice_d": "[7.5, 9.12]", + "answer_gt": "[0.0, 6.93]", + "answer_gt_word": "a car engine is idling deeply", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11269 + }, + { + "path": "Y880IpGRqC5c.wav", + "question": "How are the complete segments of 'is revved up' timed in this audio?", + "choice_a": "[0.01, 4.07]", + "choice_b": "[7.35, 9.31]", + "choice_c": "[4.04, 7.35]", + "choice_d": "[7.5, 9.31]", + "answer_gt": "[4.04, 7.35]", + "answer_gt_word": "is revved up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11270 + }, + { + "path": "Y8RjytDXc23M.wav", + "question": "At what timestamps does the complete 'a woman speaks' occur?", + "choice_a": "[1.82, 2.56]", + "choice_b": "[5.06, 5.61]", + "choice_c": "[7.64, 8.19]", + "choice_d": "[0.76, 9.72]", + "answer_gt": "[1.82, 2.56]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11271 + }, + { + "path": "Y8SYX5jXLm3A.wav", + "question": "How long does 'a hooked animal is walking in water' last each time it appears in the audio?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.22, 7.61]", + "choice_c": "[3.04, 9.26]", + "choice_d": "[3.55, 9.69]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a hooked animal is walking in water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11272 + }, + { + "path": "Y8VoiWkzvNQ4.wav", + "question": "Where does 'a mid-size motor vehicle engine is operating' appear from beginning to end in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.2, 7.83]", + "choice_c": "[5.44, 8.24]", + "choice_d": "[6.69, 9.89]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a mid-size motor vehicle engine is operating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11273 + }, + { + "path": "Y8YDpaKb3QZk.wav", + "question": "At what points does 'a car races at high velocity' exist completely in the timeline?", + "choice_a": "[2.98, 10.01]", + "choice_b": "[1.01, 8.32]", + "choice_c": "[4.23, 9.24]", + "choice_d": "[6.29, 9.41]", + "answer_gt": "[2.98, 10.01]", + "answer_gt_word": "a car races at high velocity", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11274 + }, + { + "path": "Y8ZqX99afNqE.wav", + "question": "What are the full time segments when 'a motor vehicle engine is started' appears in this sound?", + "choice_a": "[0.05, 1.84]", + "choice_b": "[8.77, 9.99]", + "choice_c": "[1.82, 8.77]", + "choice_d": "[3.74, 5.57]", + "answer_gt": "[0.05, 1.84]", + "answer_gt_word": "a motor vehicle engine is started", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11275 + }, + { + "path": "Y8aDTJApiNrk.wav", + "question": "How long does 'a vehicle car engine idling in mid frequency' last each time it appears in the audio?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[2.04, 5.88]", + "choice_c": "[2.63, 7.11]", + "choice_d": "[5.52, 9.73]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a vehicle car engine idling in mid frequency", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11276 + }, + { + "path": "Y8qAuo8Dmr2Y.wav", + "question": "At what time points is 'a person whistles' fully present in the recording?", + "choice_a": "[1.73, 3.15]", + "choice_b": "[4.02, 6.14]", + "choice_c": "[8.39, 9.98]", + "choice_d": "[9.3, 10.0]", + "answer_gt": "[9.3, 10.0]", + "answer_gt_word": "a person whistles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11277 + }, + { + "path": "Y8t4PWuA8sSM.wav", + "question": "Where is 'strong wind is blowing' consistently heard throughout the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.86, 5.75]", + "choice_c": "[4.63, 7.51]", + "choice_d": "[5.44, 9.47]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "strong wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11278 + }, + { + "path": "Y8tC9BD6s-So.wav", + "question": "What are the intervals in seconds of each 'strong wind is blowing' appearance?", + "choice_a": "[0.0, 2.01]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[9.07, 9.99]", + "choice_d": "[2.0, 5.8]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "strong wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11279 + }, + { + "path": "Y8ycflE3dIHw.wav", + "question": "Define the time intervals that fully contain 'a train passes by' in this sound.", + "choice_a": "[0.0, 4.4]", + "choice_b": "[5.84, 7.16]", + "choice_c": "[1.27, 3.56]", + "choice_d": "[2.3, 4.18]", + "answer_gt": "[0.0, 4.4]", + "answer_gt_word": "a train passes by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11280 + }, + { + "path": "Y9IshV__iHbE.wav", + "question": "Where do we hear 'wind blowing hard' in full within this sound recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.14, 3.94]", + "choice_d": "[1.42, 8.17]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind blowing hard", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11281 + }, + { + "path": "Y9lO8Hii2Q1I.wav", + "question": "In what time spans is 'train engine chugging' completely audible in this recording?", + "choice_a": "[0.0, 1.12]", + "choice_b": "[2.11, 2.59]", + "choice_c": "[3.74, 8.32]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "train engine chugging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11282 + }, + { + "path": "Y9ucfsaBXB2Q.wav", + "question": "Identify the occurrences of 'a woman laughs' along with their time frames.", + "choice_a": "[0.04, 0.42]", + "choice_b": "[1.43, 2.29]", + "choice_c": "[3.9, 4.52]", + "choice_d": "[5.07, 9.14]", + "answer_gt": "[5.07, 9.14]", + "answer_gt_word": "a woman laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11283 + }, + { + "path": "Y9wGJydlJvrk.wav", + "question": "Capture the timeline where 'engine revving rapidly' is fully included in the sound.", + "choice_a": "[0.01, 4.64]", + "choice_b": "[1.93, 5.69]", + "choice_c": "[4.4, 6.12]", + "choice_d": "[5.65, 6.79]", + "answer_gt": "[0.01, 4.64]", + "answer_gt_word": "engine revving rapidly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11284 + }, + { + "path": "YAgG_nhUurY4.wav", + "question": "What are the comprehensive timeframes of 'a motorcycle engine revs' in this clip?", + "choice_a": "[0.0, 4.25]", + "choice_b": "[1.47, 2.54]", + "choice_c": "[1.56, 3.84]", + "choice_d": "[1.96, 6.93]", + "answer_gt": "[0.0, 4.25]", + "answer_gt_word": "a motorcycle engine revs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11285 + }, + { + "path": "YAlPTUAK6ero.wav", + "question": "At what seconds do the full iterations of 'an engine humming' start and end?", + "choice_a": "[0.0, 9.02]", + "choice_b": "[3.47, 7.53]", + "choice_c": "[4.13, 7.96]", + "choice_d": "[5.37, 9.32]", + "answer_gt": "[0.0, 9.02]", + "answer_gt_word": "an engine humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11286 + }, + { + "path": "YAw89ZmdNpYU.wav", + "question": "What is the range in seconds for each full 'an engine humming' in the recording?", + "choice_a": "[0.0, 0.54]", + "choice_b": "[0.0, 9.02]", + "choice_c": "[4.13, 5.51]", + "choice_d": "[6.2, 7.53]", + "answer_gt": "[0.0, 9.02]", + "answer_gt_word": "an engine humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11287 + }, + { + "path": "YB35iQ9_XLhk.wav", + "question": "How many seconds does each full presence of 'an engine running' cover?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.12, 4.79]", + "choice_c": "[7.59, 10.01]", + "choice_d": "[7.67, 8.03]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11288 + }, + { + "path": "YB8VhLKlh9pU.wav", + "question": "At which time markers does 'wind blowing' fully unfold in the sound?", + "choice_a": "[0.0, 5.45]", + "choice_b": "[2.96, 6.84]", + "choice_c": "[3.77, 7.37]", + "choice_d": "[6.46, 8.72]", + "answer_gt": "[0.0, 5.45]", + "answer_gt_word": "wind blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11289 + }, + { + "path": "YBA-lFjpzad4.wav", + "question": "What are the comprehensive timeframes of 'a motorcycle engine drives by' in this clip?", + "choice_a": "[1.79, 10.0]", + "choice_b": "[3.32, 6.02]", + "choice_c": "[7.49, 10.0]", + "choice_d": "[6.2, 8.85]", + "answer_gt": "[1.79, 10.0]", + "answer_gt_word": "a motorcycle engine drives by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11290 + }, + { + "path": "YBA-lFjpzad4.wav", + "question": "At what time points is 'downshifting' fully present in the recording?", + "choice_a": "[6.0, 7.42]", + "choice_b": "[4.69, 6.77]", + "choice_c": "[5.99, 7.95]", + "choice_d": "[6.07, 9.32]", + "answer_gt": "[6.0, 7.42]", + "answer_gt_word": "downshifting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11291 + }, + { + "path": "YBBm27mZtcng.wav", + "question": "Determine the full durations of 'horn' in the audio file.", + "choice_a": "[1.62, 2.19]", + "choice_b": "[3.5, 10.0]", + "choice_c": "[2.38, 3.38]", + "choice_d": "[7.59, 10.0]", + "answer_gt": "[1.62, 2.19]", + "answer_gt_word": "horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11292 + }, + { + "path": "YBCrS7bCWPXs.wav", + "question": "What times mark the complete cycles of 'a very loud vibration' in the audio?", + "choice_a": "[0.0, 1.69]", + "choice_b": "[2.71, 3.27]", + "choice_c": "[4.38, 4.7]", + "choice_d": "[5.72, 6.49]", + "answer_gt": "[0.0, 1.69]", + "answer_gt_word": "a very loud vibration", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11293 + }, + { + "path": "YBCxuARguSBg.wav", + "question": "Identify the periods when 'a horn and accelerated train movement' fully resonates in the sound.", + "choice_a": "[0.01, 4.85]", + "choice_b": "[4.5, 8.8]", + "choice_c": "[6.29, 8.86]", + "choice_d": "[7.7, 9.43]", + "answer_gt": "[0.01, 4.85]", + "answer_gt_word": "a horn and accelerated train movement", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11294 + }, + { + "path": "YBFHGv_6hU0Y.wav", + "question": "Which sections of the sound contain 'engine revving loudly' in their entirety?", + "choice_a": "[0.0, 4.74]", + "choice_b": "[4.62, 10.02]", + "choice_c": "[1.05, 3.62]", + "choice_d": "[1.93, 4.83]", + "answer_gt": "[0.0, 4.74]", + "answer_gt_word": "engine revving loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11295 + }, + { + "path": "YBFgEYIWrooc.wav", + "question": "Which sections of the sound contain 'honk' in their entirety?", + "choice_a": "[0.0, 6.27]", + "choice_b": "[7.3, 10.0]", + "choice_c": "[6.09, 7.29]", + "choice_d": "[0.64, 1.94]", + "answer_gt": "[6.09, 7.29]", + "answer_gt_word": "honk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11296 + }, + { + "path": "YBHdi8AH1McE.wav", + "question": "Identify the occurrences of 'dog barking near crowd' along with their time frames.", + "choice_a": "[0.69, 9.78]", + "choice_b": "[0.64, 3.56]", + "choice_c": "[1.3, 3.57]", + "choice_d": "[2.6, 5.55]", + "answer_gt": "[0.69, 9.78]", + "answer_gt_word": "dog barking near crowd", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11297 + }, + { + "path": "YBLi6DZmr_fU.wav", + "question": "How long in seconds does 'puppies are whining' last from start to end?", + "choice_a": "[1.01, 10.02]", + "choice_b": "[0.13, 7.35]", + "choice_c": "[1.35, 8.95]", + "choice_d": "[2.14, 9.88]", + "answer_gt": "[1.01, 10.02]", + "answer_gt_word": "puppies are whining", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11298 + }, + { + "path": "YBbbCArC3CC0.wav", + "question": "What are the boundaries in seconds for each 'a woman shouts' presence?", + "choice_a": "[0.0, 0.97]", + "choice_b": "[2.85, 3.97]", + "choice_c": "[8.65, 9.99]", + "choice_d": "[1.43, 7.59]", + "answer_gt": "[8.65, 9.99]", + "answer_gt_word": "a woman shouts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11299 + }, + { + "path": "YC0AY0FEhOxU.wav", + "question": "How long in seconds does 'a cat screaming' last from start to end?", + "choice_a": "[0.0, 0.61]", + "choice_b": "[2.2, 5.73]", + "choice_c": "[6.82, 9.82]", + "choice_d": "[3.62, 3.98]", + "answer_gt": "[0.0, 0.61]", + "answer_gt_word": "a cat screaming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11300 + }, + { + "path": "YCbe2B6ohBpw.wav", + "question": "Capture the timeline where 'continuous quacking' is fully included in the sound.", + "choice_a": "[0.0, 9.02]", + "choice_b": "[0.0, 1.14]", + "choice_c": "[1.91, 2.24]", + "choice_d": "[4.29, 5.55]", + "answer_gt": "[0.0, 9.02]", + "answer_gt_word": "continuous quacking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11301 + }, + { + "path": "YChNWetbt6Xo.wav", + "question": "Locate every segment where 'an idling engine hums' is completely audible.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 0.85]", + "choice_c": "[3.48, 4.53]", + "choice_d": "[3.54, 8.03]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an idling engine hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11302 + }, + { + "path": "YD1NopL9cFPQ.wav", + "question": "At what seconds do the full iterations of 'a train horn blowing' start and end?", + "choice_a": "[0.0, 0.69]", + "choice_b": "[0.68, 10.0]", + "choice_c": "[4.46, 5.91]", + "choice_d": "[7.71, 9.27]", + "answer_gt": "[0.0, 0.69]", + "answer_gt_word": "a train horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11303 + }, + { + "path": "YD2Xc_jZllDY.wav", + "question": "Define the time intervals that fully contain 'a train horn blowing' in this sound.", + "choice_a": "[2.44, 6.18]", + "choice_b": "[0.0, 0.69]", + "choice_c": "[9.33, 9.91]", + "choice_d": "[0.87, 1.17]", + "answer_gt": "[0.0, 0.69]", + "answer_gt_word": "a train horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11304 + }, + { + "path": "YDO_ZDe2CN5I.wav", + "question": "Where do we hear 'a train horn blowing' in full within this sound recording?", + "choice_a": "[0.49, 2.4]", + "choice_b": "[3.05, 3.82]", + "choice_c": "[0.0, 0.69]", + "choice_d": "[5.46, 9.41]", + "answer_gt": "[0.0, 0.69]", + "answer_gt_word": "a train horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11305 + }, + { + "path": "YDRSiR-dkmU8.wav", + "question": "How many seconds does each full presence of 'a train horn blowing' cover?", + "choice_a": "[0.0, 0.69]", + "choice_b": "[1.69, 3.32]", + "choice_c": "[5.38, 6.44]", + "choice_d": "[7.53, 7.9]", + "answer_gt": "[0.0, 0.69]", + "answer_gt_word": "a train horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11306 + }, + { + "path": "YDcGnX0xLRLM.wav", + "question": "What are the precise second intervals of 'the horn honks a few more times' in this clip?", + "choice_a": "[0.03, 1.17]", + "choice_b": "[5.47, 7.11]", + "choice_c": "[5.5, 7.12]", + "choice_d": "[2.07, 4.67]", + "answer_gt": "[5.5, 7.12]", + "answer_gt_word": "the horn honks a few more times", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11307 + }, + { + "path": "YDx7q7vzVFGE.wav", + "question": "Determine the full durations of 'strong wind noise' in the audio file.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.13, 4.58]", + "choice_c": "[1.25, 6.25]", + "choice_d": "[3.88, 7.68]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "strong wind noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11308 + }, + { + "path": "YE70Rf1Aol3k.wav", + "question": "What are the full time segments when 'train cars click' appears in this sound?", + "choice_a": "[0.01, 10.0]", + "choice_b": "[0.01, 10.02]", + "choice_c": "[0.01, 10.02]", + "choice_d": "[9.73, 10.0]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "train cars click", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11309 + }, + { + "path": "YEQV_KGRWvgg.wav", + "question": "Identify the complete instances of 'loud humming of an aircraft engine' within the sound timeline.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.42, 5.59]", + "choice_c": "[2.77, 6.08]", + "choice_d": "[3.34, 9.51]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud humming of an aircraft engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11310 + }, + { + "path": "YE_Tqm4Twvf0.wav", + "question": "Where do we hear 'a vehicle engine loudly accelerating by' in full within this sound recording?", + "choice_a": "[0.0, 7.62]", + "choice_b": "[7.64, 10.0]", + "choice_c": "[8.62, 9.83]", + "choice_d": "[7.11, 10.0]", + "answer_gt": "[0.0, 7.62]", + "answer_gt_word": "a vehicle engine loudly accelerating by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11311 + }, + { + "path": "YEyjStn4c9lE.wav", + "question": "Where does 'a vehicle accelerates' appear from beginning to end in the sound?", + "choice_a": "[0.0, 1.59]", + "choice_b": "[7.65, 10.0]", + "choice_c": "[1.61, 7.0]", + "choice_d": "[5.1, 9.49]", + "answer_gt": "[0.0, 1.59]", + "answer_gt_word": "a vehicle accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11312 + }, + { + "path": "YFZuQBXARPLQ.wav", + "question": "Document the full appearances of 'vibrating' in terms of seconds.", + "choice_a": "[0.0, 1.6]", + "choice_b": "[9.27, 9.73]", + "choice_c": "[1.72, 7.5]", + "choice_d": "[4.53, 8.33]", + "answer_gt": "[0.0, 1.6]", + "answer_gt_word": "vibrating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11313 + }, + { + "path": "YFjtO-xQD_To.wav", + "question": "Identify the occurrences of 'speech' along with their time frames.", + "choice_a": "[0.39, 0.93]", + "choice_b": "[1.97, 3.14]", + "choice_c": "[3.14, 9.18]", + "choice_d": "[6.54, 6.92]", + "answer_gt": "[3.14, 9.18]", + "answer_gt_word": "speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11314 + }, + { + "path": "YGFQnh84kNwU.wav", + "question": "At what points does 'crossing signals ring' exist completely in the timeline?", + "choice_a": "[0.01, 3.24]", + "choice_b": "[7.72, 10.0]", + "choice_c": "[3.34, 5.46]", + "choice_d": "[7.71, 10.0]", + "answer_gt": "[3.34, 5.46]", + "answer_gt_word": "crossing signals ring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11315 + }, + { + "path": "YGJbfVBLsuQA.wav", + "question": "At what seconds do the full iterations of 'heavy rain' start and end?", + "choice_a": "[1.1, 2.25]", + "choice_b": "[7.57, 10.0]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[2.26, 5.17]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy rain", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11316 + }, + { + "path": "YHEg1bmaRqlA.wav", + "question": "Provide the start and end times of 'a siren whales' within this audio.", + "choice_a": "[0.02, 10.01]", + "choice_b": "[0.36, 2.48]", + "choice_c": "[0.64, 2.86]", + "choice_d": "[2.05, 8.11]", + "answer_gt": "[0.02, 10.01]", + "answer_gt_word": "a siren whales", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11317 + }, + { + "path": "YHqnSyliKTKA.wav", + "question": "What are the intervals in seconds of each 'an engine' appearance?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[7.19, 8.81]", + "choice_d": "[5.99, 6.32]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11318 + }, + { + "path": "YIPfaRF76gVU.wav", + "question": "What are the distinct time ranges of 'a siren is wailing' within the audio?", + "choice_a": "[0.0, 5.78]", + "choice_b": "[0.32, 1.26]", + "choice_c": "[8.28, 9.5]", + "choice_d": "[0.01, 9.5]", + "answer_gt": "[0.0, 5.78]", + "answer_gt_word": "a siren is wailing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11319 + }, + { + "path": "YIclFv7GGoDI.wav", + "question": "What timestamps mark the full occurrences of 'a siren wails'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.13, 3.2]", + "choice_c": "[6.26, 9.18]", + "choice_d": "[1.44, 4.1]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a siren wails", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11320 + }, + { + "path": "YJ8Npt2M7tzE.wav", + "question": "Capture the timeline where 'a vehicle horns' is fully included in the sound.", + "choice_a": "[0.44, 1.11]", + "choice_b": "[1.41, 2.17]", + "choice_c": "[2.47, 3.32]", + "choice_d": "[4.19, 5.01]", + "answer_gt": "[0.44, 1.11]", + "answer_gt_word": "a vehicle horns", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11321 + }, + { + "path": "YK6-Xm6FnX8s.wav", + "question": "Determine the full durations of 'horns blow from a humming oncoming train' in the audio file.", + "choice_a": "[0.0, 8.09]", + "choice_b": "[2.15, 5.44]", + "choice_c": "[2.73, 8.03]", + "choice_d": "[5.19, 8.15]", + "answer_gt": "[0.0, 8.09]", + "answer_gt_word": "horns blow from a humming oncoming train", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11322 + }, + { + "path": "YK9lmiWhXu1s.wav", + "question": "How long does 'an engine' last each time it appears in the audio?", + "choice_a": "[1.02, 5.21]", + "choice_b": "[2.19, 5.24]", + "choice_c": "[5.26, 8.01]", + "choice_d": "[1.48, 9.33]", + "answer_gt": "[1.02, 5.21]", + "answer_gt_word": "an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11323 + }, + { + "path": "YKniufpODcOg.wav", + "question": "What are the precise second intervals of 'an engine' in this clip?", + "choice_a": "[0.02, 0.46]", + "choice_b": "[1.56, 3.92]", + "choice_c": "[4.78, 8.32]", + "choice_d": "[1.02, 5.21]", + "answer_gt": "[1.02, 5.21]", + "answer_gt_word": "an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11324 + }, + { + "path": "YLBpmR72vQBk.wav", + "question": "How long does 'a mid-size motor vehicle engine idles' last each time it appears in the audio?", + "choice_a": "[7.83, 10.0]", + "choice_b": "[0.0, 5.2]", + "choice_c": "[3.37, 4.02]", + "choice_d": "[5.03, 5.24]", + "answer_gt": "[7.83, 10.0]", + "answer_gt_word": "a mid-size motor vehicle engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11325 + }, + { + "path": "YLbJZuD_Y_uM.wav", + "question": "What second marks correspond to the entire span of 'a steady wind blows'?", + "choice_a": "[0.01, 10.02]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.0, 1.04]", + "choice_d": "[1.24, 1.83]", + "answer_gt": "[0.01, 10.02]", + "answer_gt_word": "a steady wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11326 + }, + { + "path": "YLyeZhAu7yh4.wav", + "question": "What are the precise second intervals of 'a man talks' in this clip?", + "choice_a": "[0.83, 2.52]", + "choice_b": "[4.33, 5.19]", + "choice_c": "[5.62, 6.62]", + "choice_d": "[2.65, 4.21]", + "answer_gt": "[5.62, 6.62]", + "answer_gt_word": "a man talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11327 + }, + { + "path": "YLzcNa3HvD7c.wav", + "question": "How are the complete segments of 'engine rumbling' timed in this audio?", + "choice_a": "[0.0, 1.02]", + "choice_b": "[1.8, 2.28]", + "choice_c": "[3.53, 4.24]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "engine rumbling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11328 + }, + { + "path": "YM7Ba3iTY0rE.wav", + "question": "At which seconds does 'honk' fully register in the sound?", + "choice_a": "[0.0, 3.56]", + "choice_b": "[1.34, 3.02]", + "choice_c": "[2.5, 8.85]", + "choice_d": "[2.58, 9.22]", + "answer_gt": "[0.0, 3.56]", + "answer_gt_word": "honk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11329 + }, + { + "path": "YMbmDn7ahGEg.wav", + "question": "What times mark the complete cycles of 'white noise together' in the audio?", + "choice_a": "[0.0, 9.01]", + "choice_b": "[2.23, 7.53]", + "choice_c": "[8.47, 9.01]", + "choice_d": "[2.11, 7.05]", + "answer_gt": "[0.0, 9.01]", + "answer_gt_word": "white noise together", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11330 + }, + { + "path": "YMvHpNzDpC6Q.wav", + "question": "What are the precise second intervals of 'white noise together' in this clip?", + "choice_a": "[0.4, 1.54]", + "choice_b": "[2.59, 2.93]", + "choice_c": "[4.02, 4.49]", + "choice_d": "[0.0, 9.01]", + "answer_gt": "[0.0, 9.01]", + "answer_gt_word": "white noise together", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11331 + }, + { + "path": "YNDaVSIJaXVs.wav", + "question": "What are the comprehensive timeframes of 'an aircraft engine operates' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.48, 3.39]", + "choice_c": "[1.46, 4.19]", + "choice_d": "[3.22, 8.05]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an aircraft engine operates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11332 + }, + { + "path": "YNeWW30WZjPc.wav", + "question": "In what time spans is 'an aircraft engine operates' completely audible in this recording?", + "choice_a": "[0.0, 0.78]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[3.48, 3.88]", + "choice_d": "[8.82, 9.23]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an aircraft engine operates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11333 + }, + { + "path": "YNeuFvAzcDaI.wav", + "question": "At which time markers does 'the sound of an aircraft' fully unfold in the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[3.93, 8.7]", + "choice_c": "[4.1, 8.74]", + "choice_d": "[4.84, 9.27]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "the sound of an aircraft", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11334 + }, + { + "path": "YO5wKqu7fcyw.wav", + "question": "Where do we hear 'several emergency sirens are triggered' in full within this sound recording?", + "choice_a": "[0.0, 3.84]", + "choice_b": "[0.24, 5.51]", + "choice_c": "[2.65, 9.29]", + "choice_d": "[3.74, 9.36]", + "answer_gt": "[0.0, 3.84]", + "answer_gt_word": "several emergency sirens are triggered", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11335 + }, + { + "path": "YOaQR4KDY1Ug.wav", + "question": "What timestamps mark the full occurrences of 'a vehicle accelerates loudly down the road'?", + "choice_a": "[3.41, 9.76]", + "choice_b": "[0.86, 4.74]", + "choice_c": "[1.0, 6.15]", + "choice_d": "[3.66, 7.3]", + "answer_gt": "[3.41, 9.76]", + "answer_gt_word": "a vehicle accelerates loudly down the road", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11336 + }, + { + "path": "YP808463iRRU.wav", + "question": "At what seconds do the full iterations of 'a music is played' start and end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[6.56, 7.57]", + "choice_c": "[8.96, 9.49]", + "choice_d": "[0.0, 0.78]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a music is played", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11337 + }, + { + "path": "YPRBAm39xujU.wav", + "question": "At which time markers does 'a mid-size motor vehicle engine idles' fully unfold in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[4.59, 7.08]", + "choice_d": "[5.88, 9.98]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a mid-size motor vehicle engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11338 + }, + { + "path": "YQBLpskcb_As.wav", + "question": "How are the complete segments of 'a loud train whistle' timed in this audio?", + "choice_a": "[0.35, 2.52]", + "choice_b": "[0.78, 3.14]", + "choice_c": "[1.84, 6.62]", + "choice_d": "[3.12, 7.58]", + "answer_gt": "[0.35, 2.52]", + "answer_gt_word": "a loud train whistle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11339 + }, + { + "path": "YQHVho4OnVtc.wav", + "question": "Locate every segment where 'a train horn sounds' is completely audible.", + "choice_a": "[2.42, 5.68]", + "choice_b": "[0.26, 5.44]", + "choice_c": "[1.16, 6.79]", + "choice_d": "[4.3, 8.33]", + "answer_gt": "[2.42, 5.68]", + "answer_gt_word": "a train horn sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11340 + }, + { + "path": "YQWWAj4WZuyU.wav", + "question": "At what time intervals does 'a train horn sounds' occur in the sound file?", + "choice_a": "[0.12, 4.41]", + "choice_b": "[7.3, 9.98]", + "choice_c": "[0.14, 1.03]", + "choice_d": "[2.42, 5.68]", + "answer_gt": "[2.42, 5.68]", + "answer_gt_word": "a train horn sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11341 + }, + { + "path": "YQa8QRHe31m8.wav", + "question": "Document the full appearances of 'a man speaks' in terms of seconds.", + "choice_a": "[1.08, 1.51]", + "choice_b": "[1.69, 2.86]", + "choice_c": "[4.87, 5.88]", + "choice_d": "[7.42, 8.62]", + "answer_gt": "[1.08, 1.51]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11342 + }, + { + "path": "YQvATUKXYFBs.wav", + "question": "What are the exact times when 'railroad warning signals ring' is present in the clip?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.73, 1.31]", + "choice_c": "[2.28, 3.11]", + "choice_d": "[7.97, 8.77]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "railroad warning signals ring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11343 + }, + { + "path": "YRGxGn-VGW5s.wav", + "question": "In what time spans is 'an adult male shouts once in the background' completely audible in this recording?", + "choice_a": "[0.0, 0.95]", + "choice_b": "[6.22, 10.0]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[0.0, 6.29]", + "answer_gt": "[0.0, 0.95]", + "answer_gt_word": "an adult male shouts once in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11344 + }, + { + "path": "YRJrJAyUZwc4.wav", + "question": "What time segments showcase 'dogs bark' in its entirety?", + "choice_a": "[0.0, 7.26]", + "choice_b": "[1.16, 1.58]", + "choice_c": "[2.71, 3.08]", + "choice_d": "[4.03, 4.28]", + "answer_gt": "[0.0, 7.26]", + "answer_gt_word": "dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11345 + }, + { + "path": "YRLYc5lfS3eE.wav", + "question": "What are the precise second intervals of 'ambulance emergency siren' in this clip?", + "choice_a": "[0.0, 9.76]", + "choice_b": "[0.46, 8.48]", + "choice_c": "[2.86, 8.9]", + "choice_d": "[3.1, 9.48]", + "answer_gt": "[0.0, 9.76]", + "answer_gt_word": "ambulance emergency siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11346 + }, + { + "path": "YRPvqI4KYiO4.wav", + "question": "What are the specific start and stop times of 'a car beeps loudly' in the sound?", + "choice_a": "[1.2, 6.34]", + "choice_b": "[4.6, 10.0]", + "choice_c": "[2.96, 5.06]", + "choice_d": "[3.16, 7.32]", + "answer_gt": "[1.2, 6.34]", + "answer_gt_word": "a car beeps loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11347 + }, + { + "path": "YRZZKDnECWyQ.wav", + "question": "What second marks correspond to the entire span of 'sounding the siren'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.65, 8.4]", + "choice_c": "[3.8, 8.51]", + "choice_d": "[5.49, 8.76]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sounding the siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11348 + }, + { + "path": "YRmXHbnsKe0g.wav", + "question": "In what segments can 'wind whips by violently' be heard from start to finish?", + "choice_a": "[0.0, 6.37]", + "choice_b": "[0.04, 2.52]", + "choice_c": "[0.68, 6.76]", + "choice_d": "[2.1, 7.96]", + "answer_gt": "[0.0, 6.37]", + "answer_gt_word": "wind whips by violently", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11349 + }, + { + "path": "YSTIzm2JFbuc.wav", + "question": "What are the comprehensive timeframes of 'clickety-clack sounds' in this clip?", + "choice_a": "[2.4, 7.22]", + "choice_b": "[0.79, 2.09]", + "choice_c": "[0.45, 4.25]", + "choice_d": "[3.02, 4.98]", + "answer_gt": "[2.4, 7.22]", + "answer_gt_word": "clickety-clack sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11350 + }, + { + "path": "YSrobrl1Lh1A.wav", + "question": "Capture the timeline where 'wind blows' is fully included in the sound.", + "choice_a": "[0.0, 10.01]", + "choice_b": "[3.89, 4.34]", + "choice_c": "[8.34, 8.78]", + "choice_d": "[0.26, 7.0]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11351 + }, + { + "path": "YSz7po5sNgQw.wav", + "question": "Provide the start and end times of 'wind blows' within this audio.", + "choice_a": "[0.0, 2.03]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[5.61, 7.31]", + "choice_d": "[8.82, 10.0]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11352 + }, + { + "path": "YUR-fBo0fmRo.wav", + "question": "Capture the timeline where 'a vehicle engine revving and accelerating' is fully included in the sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[2.25, 10.0]", + "choice_c": "[2.38, 9.98]", + "choice_d": "[4.82, 5.13]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a vehicle engine revving and accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11353 + }, + { + "path": "YUW8czMvlmK4.wav", + "question": "What ranges in seconds capture 'booming and vibrating engines revving' in full?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[5.11, 7.84]", + "choice_c": "[5.24, 9.79]", + "choice_d": "[5.65, 9.83]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "booming and vibrating engines revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11354 + }, + { + "path": "YVb9krAYFfWE.wav", + "question": "Document the full appearances of 'a very close train engine' in terms of seconds.", + "choice_a": "[0.0, 6.64]", + "choice_b": "[0.01, 1.7]", + "choice_c": "[2.91, 6.6]", + "choice_d": "[1.22, 6.24]", + "answer_gt": "[0.0, 6.64]", + "answer_gt_word": "a very close train engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11355 + }, + { + "path": "YVrJlserGqxI.wav", + "question": "At which time markers does 'honk' fully unfold in the sound?", + "choice_a": "[1.65, 2.82]", + "choice_b": "[4.49, 5.79]", + "choice_c": "[6.54, 9.46]", + "choice_d": "[5.77, 6.21]", + "answer_gt": "[5.77, 6.21]", + "answer_gt_word": "honk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11356 + }, + { + "path": "YXG5f9OyzO_A.wav", + "question": "Capture the timeline where 'an engine accelerating' is fully included in the sound.", + "choice_a": "[0.0, 10.01]", + "choice_b": "[0.52, 1.5]", + "choice_c": "[2.65, 5.49]", + "choice_d": "[6.3, 10.01]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an engine accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11357 + }, + { + "path": "YXL8JV9qXGLE.wav", + "question": "How long in seconds does 'an engine accelerating' last from start to end?", + "choice_a": "[0.76, 2.47]", + "choice_b": "[3.77, 4.02]", + "choice_c": "[5.36, 9.18]", + "choice_d": "[0.0, 10.01]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an engine accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11358 + }, + { + "path": "YXi6V0LGvqoo.wav", + "question": "Determine the full durations of 'a dog whimpering' in the audio file.", + "choice_a": "[0.44, 1.86]", + "choice_b": "[3.55, 6.42]", + "choice_c": "[7.93, 8.71]", + "choice_d": "[6.42, 7.64]", + "answer_gt": "[0.44, 1.86]", + "answer_gt_word": "a dog whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11359 + }, + { + "path": "YXi6V0LGvqoo.wav", + "question": "At what time points is 'whine' fully present in the recording?", + "choice_a": "[1.13, 2.02]", + "choice_b": "[0.03, 4.79]", + "choice_c": "[0.29, 8.63]", + "choice_d": "[4.31, 9.16]", + "answer_gt": "[1.13, 2.02]", + "answer_gt_word": "whine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11360 + }, + { + "path": "YZG4byeZq9S4.wav", + "question": "How are the complete segments of 'an engine is idling' timed in this audio?", + "choice_a": "[0.12, 10.0]", + "choice_b": "[0.0, 0.12]", + "choice_c": "[3.22, 9.2]", + "choice_d": "[6.4, 9.68]", + "answer_gt": "[0.12, 10.0]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11361 + }, + { + "path": "YZK1kWFoHLQo.wav", + "question": "What are the boundaries in seconds for each 'a cat meows' presence?", + "choice_a": "[0.17, 1.08]", + "choice_b": "[8.64, 10.01]", + "choice_c": "[6.05, 7.23]", + "choice_d": "[8.02, 8.37]", + "answer_gt": "[8.64, 10.01]", + "answer_gt_word": "a cat meows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11362 + }, + { + "path": "Y_5DSfZnQZPw.wav", + "question": "What are the full time segments when 'people laughing' appears in this sound?", + "choice_a": "[0.61, 1.24]", + "choice_b": "[2.68, 2.88]", + "choice_c": "[4.22, 5.04]", + "choice_d": "[8.73, 9.71]", + "answer_gt": "[4.22, 5.04]", + "answer_gt_word": "people laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11363 + }, + { + "path": "Y_SHrBNEes38.wav", + "question": "List the specific seconds of 'people laughing' appearances.", + "choice_a": "[4.22, 5.04]", + "choice_b": "[2.0, 2.27]", + "choice_c": "[2.81, 3.11]", + "choice_d": "[3.63, 3.94]", + "answer_gt": "[4.22, 5.04]", + "answer_gt_word": "people laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11364 + }, + { + "path": "Y__a3v2k4Svg.wav", + "question": "List the specific seconds of 'dogs whimper loudly nearby several times' appearances.", + "choice_a": "[5.8, 7.35]", + "choice_b": "[0.91, 9.13]", + "choice_c": "[5.62, 9.2]", + "choice_d": "[7.3, 9.71]", + "answer_gt": "[5.8, 7.35]", + "answer_gt_word": "dogs whimper loudly nearby several times", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11365 + }, + { + "path": "Y_kL6LQF41cA.wav", + "question": "Document the full appearances of 'wind blows' in terms of seconds.", + "choice_a": "[4.58, 10.0]", + "choice_b": "[1.0, 2.24]", + "choice_c": "[1.22, 4.34]", + "choice_d": "[1.74, 9.74]", + "answer_gt": "[4.58, 10.0]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11366 + }, + { + "path": "YaWhMvg9JbxM.wav", + "question": "What are the intervals in seconds of each 'dogs bark' appearance?", + "choice_a": "[0.0, 9.87]", + "choice_b": "[0.07, 0.32]", + "choice_c": "[2.14, 2.54]", + "choice_d": "[5.5, 8.59]", + "answer_gt": "[0.0, 9.87]", + "answer_gt_word": "dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11367 + }, + { + "path": "YayUqgnURxgo.wav", + "question": "Where is 'wind blows loudly nearby' found in its entirety within the audio's timeline?", + "choice_a": "[0.82, 9.41]", + "choice_b": "[1.17, 3.93]", + "choice_c": "[5.69, 6.2]", + "choice_d": "[6.95, 9.4]", + "answer_gt": "[0.82, 9.41]", + "answer_gt_word": "wind blows loudly nearby", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11368 + }, + { + "path": "YbS4flke6PUg.wav", + "question": "List the periods when 'woman talking' is completely played in the sound.", + "choice_a": "[8.16, 10.0]", + "choice_b": "[0.38, 1.3]", + "choice_c": "[5.14, 6.43]", + "choice_d": "[7.21, 8.02]", + "answer_gt": "[8.16, 10.0]", + "answer_gt_word": "woman talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11369 + }, + { + "path": "YbygBWUkpaC8.wav", + "question": "What time segments delineate the presence of 'a male speech'?", + "choice_a": "[0.0, 3.25]", + "choice_b": "[0.92, 1.34]", + "choice_c": "[2.98, 10.0]", + "choice_d": "[2.3, 2.55]", + "answer_gt": "[0.0, 3.25]", + "answer_gt_word": "a male speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11370 + }, + { + "path": "Yc3nlaAkv9bA.wav", + "question": "How are the complete segments of 'goats baaing' timed in this audio?", + "choice_a": "[0.0, 0.41]", + "choice_b": "[1.37, 2.85]", + "choice_c": "[3.13, 4.18]", + "choice_d": "[5.27, 7.34]", + "answer_gt": "[3.13, 4.18]", + "answer_gt_word": "goats baaing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11371 + }, + { + "path": "YcHm9Hdis3wE.wav", + "question": "How long in seconds does 'a motor engine is idling' last from start to end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.16, 4.61]", + "choice_c": "[0.39, 7.88]", + "choice_d": "[2.15, 9.53]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a motor engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11372 + }, + { + "path": "YcNn1rxrQq9c.wav", + "question": "What timestamps mark the full occurrences of 'a vehicle engine revs'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[4.51, 6.22]", + "choice_c": "[7.79, 9.62]", + "choice_d": "[4.3, 7.33]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a vehicle engine revs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11373 + }, + { + "path": "YcnBAAZDhuVA.wav", + "question": "What are all the complete time segments of 'a fire truck siren is blaring' in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.57, 4.22]", + "choice_c": "[1.39, 4.5]", + "choice_d": "[1.97, 6.75]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a fire truck siren is blaring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11374 + }, + { + "path": "YdAepTxzUYFQ.wav", + "question": "At which seconds does 'a woman speaking' fully register in the sound?", + "choice_a": "[0.16, 3.19]", + "choice_b": "[3.41, 10.0]", + "choice_c": "[4.32, 8.64]", + "choice_d": "[0.96, 1.48]", + "answer_gt": "[0.16, 3.19]", + "answer_gt_word": "a woman speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11375 + }, + { + "path": "YeTok2QICK6Y.wav", + "question": "Where does 'a long railroad whistle' appear from beginning to end in the sound?", + "choice_a": "[0.65, 5.15]", + "choice_b": "[9.68, 10.0]", + "choice_c": "[4.05, 9.61]", + "choice_d": "[4.42, 5.14]", + "answer_gt": "[0.65, 5.15]", + "answer_gt_word": "a long railroad whistle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11376 + }, + { + "path": "YelztUCeNQvQ.wav", + "question": "What are the boundaries in seconds for each 'a train horn honking' presence?", + "choice_a": "[0.0, 0.67]", + "choice_b": "[3.5, 6.98]", + "choice_c": "[3.65, 7.47]", + "choice_d": "[5.36, 9.44]", + "answer_gt": "[0.0, 0.67]", + "answer_gt_word": "a train horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11377 + }, + { + "path": "Yf98tFMTAOrQ.wav", + "question": "What are the specific start and stop times of 'a train horn honking' in the sound?", + "choice_a": "[0.0, 1.27]", + "choice_b": "[0.0, 0.67]", + "choice_c": "[0.97, 1.71]", + "choice_d": "[2.85, 5.57]", + "answer_gt": "[0.0, 0.67]", + "answer_gt_word": "a train horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11378 + }, + { + "path": "YfM10Whc-ro0.wav", + "question": "At what points does 'a high pitched engine buzzing' exist completely in the timeline?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.54, 2.56]", + "choice_c": "[3.92, 4.41]", + "choice_d": "[5.59, 6.47]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a high pitched engine buzzing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11379 + }, + { + "path": "Yfg_unNt1yds.wav", + "question": "What are all the complete time segments of 'racing vehicles revving and accelerating' in this sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.54, 3.85]", + "choice_c": "[3.59, 4.03]", + "choice_d": "[3.8, 7.73]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "racing vehicles revving and accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11380 + }, + { + "path": "YgXJXfZra9TA.wav", + "question": "What time segments showcase 'several people cheer' in its entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.73, 4.59]", + "choice_c": "[3.33, 4.76]", + "choice_d": "[3.69, 4.78]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "several people cheer", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11381 + }, + { + "path": "YgYDoraoD85E.wav", + "question": "At what seconds do the full iterations of 'a light aircraft engine passes nearby' start and end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[4.94, 7.25]", + "choice_c": "[5.72, 8.65]", + "choice_d": "[7.22, 9.48]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a light aircraft engine passes nearby", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11382 + }, + { + "path": "Yi2tHuOnJTWY.wav", + "question": "What ranges in seconds capture 'a vehicle downshifts' in full?", + "choice_a": "[0.17, 10.0]", + "choice_b": "[0.17, 10.0]", + "choice_c": "[0.26, 2.48]", + "choice_d": "[4.52, 4.97]", + "answer_gt": "[0.17, 10.0]", + "answer_gt_word": "a vehicle downshifts", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11383 + }, + { + "path": "YiQGztcbPgWI.wav", + "question": "Identify the occurrences of 'blows a horn' along with their time frames.", + "choice_a": "[9.32, 10.0]", + "choice_b": "[1.05, 4.92]", + "choice_c": "[1.72, 8.12]", + "choice_d": "[4.26, 9.78]", + "answer_gt": "[9.32, 10.0]", + "answer_gt_word": "blows a horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11384 + }, + { + "path": "YkPfjp_C66HA.wav", + "question": "List the specific seconds of 'clanging sounds' appearances.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.96, 2.72]", + "choice_c": "[3.3, 3.85]", + "choice_d": "[6.29, 7.09]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "clanging sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11385 + }, + { + "path": "YkY1wpO7_pMs.wav", + "question": "What are the full time segments when 'clanging sounds' appears in this sound?", + "choice_a": "[0.38, 0.85]", + "choice_b": "[4.74, 6.02]", + "choice_c": "[6.91, 10.0]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "clanging sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11386 + }, + { + "path": "Ykk_Q4naP_9o.wav", + "question": "Where do we hear 'a humming engine' in full within this sound recording?", + "choice_a": "[1.69, 2.27]", + "choice_b": "[2.83, 5.59]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[3.12, 7.7]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a humming engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11387 + }, + { + "path": "Yl96ZcCdYwfY.wav", + "question": "List the periods when 'clickety clanking' is completely played in the sound.", + "choice_a": "[0.0, 1.37]", + "choice_b": "[9.29, 10.0]", + "choice_c": "[2.36, 10.01]", + "choice_d": "[7.38, 9.27]", + "answer_gt": "[2.36, 10.01]", + "answer_gt_word": "clickety clanking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11388 + }, + { + "path": "YlDtptH9Odf4.wav", + "question": "Where is 'a man speaking after' found in its entirety within the audio's timeline?", + "choice_a": "[5.73, 7.07]", + "choice_b": "[0.21, 5.02]", + "choice_c": "[2.19, 5.62]", + "choice_d": "[3.19, 6.69]", + "answer_gt": "[5.73, 7.07]", + "answer_gt_word": "a man speaking after", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11389 + }, + { + "path": "YlemaMeWsO1s.wav", + "question": "What are the distinct time ranges of 'an engine humming' within the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.5, 8.58]", + "choice_c": "[3.76, 8.92]", + "choice_d": "[6.16, 9.22]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11390 + }, + { + "path": "Ylywxhiw977o.wav", + "question": "How long in seconds does 'an idle vehicle engine running' last from start to end?", + "choice_a": "[0.01, 10.01]", + "choice_b": "[2.78, 4.68]", + "choice_c": "[4.44, 5.2]", + "choice_d": "[4.48, 6.67]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "an idle vehicle engine running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11391 + }, + { + "path": "YmKaqk8Fc4Kc.wav", + "question": "How long in seconds does 'a vehicle engine revving' last from start to end?", + "choice_a": "[5.08, 10.0]", + "choice_b": "[4.8, 10.0]", + "choice_c": "[1.58, 3.33]", + "choice_d": "[4.11, 6.12]", + "answer_gt": "[5.08, 10.0]", + "answer_gt_word": "a vehicle engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11392 + }, + { + "path": "YmUpVz-n8FqQ.wav", + "question": "How many seconds does each full presence of 'a siren wails' cover?", + "choice_a": "[0.0, 7.79]", + "choice_b": "[7.75, 9.25]", + "choice_c": "[6.41, 8.35]", + "choice_d": "[6.73, 8.8]", + "answer_gt": "[0.0, 7.79]", + "answer_gt_word": "a siren wails", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11393 + }, + { + "path": "Yn7IFnrK8POM.wav", + "question": "What timestamps mark the full occurrences of 'a steady heavy rain is falling'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.42, 1.85]", + "choice_c": "[5.41, 6.8]", + "choice_d": "[7.39, 8.27]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a steady heavy rain is falling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11394 + }, + { + "path": "YoN0IcZaHD_8.wav", + "question": "How long does 'drilling' last each time it appears in the audio?", + "choice_a": "[0.0, 0.64]", + "choice_b": "[8.58, 10.0]", + "choice_c": "[1.22, 5.21]", + "choice_d": "[3.33, 5.65]", + "answer_gt": "[1.22, 5.21]", + "answer_gt_word": "drilling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11395 + }, + { + "path": "YoNSWAH9mRxg.wav", + "question": "Document the full appearances of 'drilling' in terms of seconds.", + "choice_a": "[0.0, 2.46]", + "choice_b": "[4.64, 7.61]", + "choice_c": "[1.22, 5.21]", + "choice_d": "[3.12, 3.45]", + "answer_gt": "[1.22, 5.21]", + "answer_gt_word": "drilling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11396 + }, + { + "path": "YpbiQmndSpoI.wav", + "question": "What are all the complete time segments of 'an aircraft engine is operating' in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.83, 5.56]", + "choice_c": "[4.31, 7.13]", + "choice_d": "[5.17, 8.36]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an aircraft engine is operating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11397 + }, + { + "path": "Ypoo0WZI_ERw.wav", + "question": "What are the intervals in seconds of each 'grass rustling' appearance?", + "choice_a": "[0.0, 3.25]", + "choice_b": "[3.62, 8.17]", + "choice_c": "[9.42, 9.73]", + "choice_d": "[2.68, 10.0]", + "answer_gt": "[2.68, 10.0]", + "answer_gt_word": "grass rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11398 + }, + { + "path": "YqHDCNzaO208.wav", + "question": "Locate every segment where 'grass rustling' is completely audible.", + "choice_a": "[0.0, 0.22]", + "choice_b": "[2.08, 3.97]", + "choice_c": "[2.68, 10.0]", + "choice_d": "[9.08, 10.0]", + "answer_gt": "[2.68, 10.0]", + "answer_gt_word": "grass rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11399 + }, + { + "path": "YqOA4vLvHHn8.wav", + "question": "In what segments can 'grass rustling' be heard from start to finish?", + "choice_a": "[2.68, 10.0]", + "choice_b": "[1.14, 3.02]", + "choice_c": "[4.74, 5.57]", + "choice_d": "[6.29, 7.61]", + "answer_gt": "[2.68, 10.0]", + "answer_gt_word": "grass rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11400 + }, + { + "path": "YriuGGhE1avc.wav", + "question": "How many seconds does each full presence of 'a man briefly speaking through a television speaker in the background' cover?", + "choice_a": "[6.02, 6.61]", + "choice_b": "[2.24, 5.91]", + "choice_c": "[6.7, 7.44]", + "choice_d": "[8.65, 10.0]", + "answer_gt": "[6.02, 6.61]", + "answer_gt_word": "a man briefly speaking through a television speaker in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11401 + }, + { + "path": "YsNSmgP4P6Uc.wav", + "question": "At what seconds do the full iterations of 'police car siren' start and end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.09, 6.35]", + "choice_c": "[2.1, 8.42]", + "choice_d": "[4.72, 9.52]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "police car siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11402 + }, + { + "path": "Ysd7x_Vssqlw.wav", + "question": "How many seconds does 'a vehicle accelerates squealing tires' span in each occurrence?", + "choice_a": "[0.0, 8.03]", + "choice_b": "[0.06, 2.25]", + "choice_c": "[1.01, 3.47]", + "choice_d": "[1.27, 4.16]", + "answer_gt": "[0.0, 8.03]", + "answer_gt_word": "a vehicle accelerates squealing tires", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11403 + }, + { + "path": "YsiVH9ZLOZnQ.wav", + "question": "What timestamps mark the full occurrences of 'cars drifting and accelerating'?", + "choice_a": "[0.06, 10.0]", + "choice_b": "[2.62, 5.58]", + "choice_c": "[3.01, 7.26]", + "choice_d": "[4.07, 8.86]", + "answer_gt": "[0.06, 10.0]", + "answer_gt_word": "cars drifting and accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11404 + }, + { + "path": "YtDVADusiIoc.wav", + "question": "What times mark the complete cycles of 'a man speaks on a radio' in the audio?", + "choice_a": "[0.98, 5.31]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.15, 1.08]", + "choice_d": "[0.57, 7.89]", + "answer_gt": "[0.98, 5.31]", + "answer_gt_word": "a man speaks on a radio", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11405 + }, + { + "path": "YuUV7Y4f0WCI.wav", + "question": "Document the full appearances of 'loud clanging' in terms of seconds.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.44, 9.09]", + "choice_c": "[4.98, 5.85]", + "choice_d": "[5.43, 8.24]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud clanging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11406 + }, + { + "path": "YuWPRNLnpy7Y.wav", + "question": "What time segments delineate the presence of 'a vehicles accelerate quickly'?", + "choice_a": "[0.0, 4.73]", + "choice_b": "[6.19, 8.43]", + "choice_c": "[0.39, 5.01]", + "choice_d": "[2.77, 7.51]", + "answer_gt": "[0.0, 4.73]", + "answer_gt_word": "a vehicles accelerate quickly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11407 + }, + { + "path": "Yulwx9Jx_hUg.wav", + "question": "Where is 'motor running' consistently heard throughout the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[6.91, 8.69]", + "choice_c": "[0.98, 7.73]", + "choice_d": "[2.52, 9.51]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "motor running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11408 + }, + { + "path": "YvNBZJqbiv0c.wav", + "question": "What are the comprehensive timeframes of 'train running blowing horn' in this clip?", + "choice_a": "[4.76, 6.46]", + "choice_b": "[1.25, 4.28]", + "choice_c": "[8.37, 10.0]", + "choice_d": "[1.94, 4.14]", + "answer_gt": "[4.76, 6.46]", + "answer_gt_word": "train running blowing horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11409 + }, + { + "path": "YvVhthZ45k3Y.wav", + "question": "Define the time intervals that fully contain 'a man laughs' in this sound.", + "choice_a": "[0.32, 0.86]", + "choice_b": "[0.0, 6.27]", + "choice_c": "[7.15, 7.58]", + "choice_d": "[8.54, 10.0]", + "answer_gt": "[0.32, 0.86]", + "answer_gt_word": "a man laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11410 + }, + { + "path": "YvVhthZ45k3Y.wav", + "question": "What are the exact times when 'hiss' is present in the clip?", + "choice_a": "[7.79, 8.56]", + "choice_b": "[0.0, 6.32]", + "choice_c": "[7.18, 7.57]", + "choice_d": "[8.54, 10.0]", + "answer_gt": "[7.79, 8.56]", + "answer_gt_word": "hiss", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11411 + }, + { + "path": "YwF3poC0J20c.wav", + "question": "What are the intervals in seconds of each 'loud music plays continuously' appearance?", + "choice_a": "[0.01, 3.04]", + "choice_b": "[3.1, 10.0]", + "choice_c": "[3.36, 7.82]", + "choice_d": "[4.21, 9.32]", + "answer_gt": "[0.01, 3.04]", + "answer_gt_word": "loud music plays continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11412 + }, + { + "path": "YwVi5w_NU6CM.wav", + "question": "What are the precise second intervals of 'emergency sirens wail' in this clip?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[1.79, 6.8]", + "choice_d": "[4.96, 8.48]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "emergency sirens wail", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11413 + }, + { + "path": "YwqZ135_Ssz0.wav", + "question": "Determine the full durations of 'bird chirps' in the audio file.", + "choice_a": "[0.0, 1.33]", + "choice_b": "[2.52, 4.72]", + "choice_c": "[6.42, 10.0]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "bird chirps", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11414 + }, + { + "path": "YxIS6JwaGxuk.wav", + "question": "Where does 'bird chirps' appear from beginning to end in the sound?", + "choice_a": "[0.0, 1.84]", + "choice_b": "[3.12, 4.13]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[2.29, 3.14]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "bird chirps", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11415 + }, + { + "path": "Yxfaoyyzw2WU.wav", + "question": "What time segments showcase 'a jet engine roars' in its entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.71, 10.0]", + "choice_c": "[4.13, 7.53]", + "choice_d": "[6.96, 8.51]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a jet engine roars", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11416 + }, + { + "path": "Yxn4AgQ3Ekrc.wav", + "question": "Which sections of the sound contain 'a jet engine roars' in their entirety?", + "choice_a": "[0.0, 1.5]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[4.44, 8.17]", + "choice_d": "[2.07, 2.5]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a jet engine roars", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11417 + }, + { + "path": "YxnVzaYbBspo.wav", + "question": "What are the intervals in seconds of each 'a dark growl' appearance?", + "choice_a": "[0.0, 9.72]", + "choice_b": "[0.0, 9.71]", + "choice_c": "[0.0, 9.72]", + "choice_d": "[4.78, 8.18]", + "answer_gt": "[0.0, 9.72]", + "answer_gt_word": "a dark growl", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11418 + }, + { + "path": "Yxx3yKzrSOBI.wav", + "question": "In what time spans is 'an emergency vehicle blares its siren' completely audible in this recording?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 0.8]", + "choice_c": "[1.84, 2.53]", + "choice_d": "[3.76, 4.26]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an emergency vehicle blares its siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11419 + }, + { + "path": "Yxz6IFHbr_Ac.wav", + "question": "What is the range in seconds for each full 'an engine roars loudly' in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.79, 2.76]", + "choice_c": "[1.86, 6.22]", + "choice_d": "[2.09, 9.66]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine roars loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11420 + }, + { + "path": "YyRUZQSU4EzQ.wav", + "question": "What are the exact times when 'an engine roars loudly' is present in the clip?", + "choice_a": "[0.03, 0.82]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[8.41, 8.78]", + "choice_d": "[9.46, 9.99]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine roars loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11421 + }, + { + "path": "YydTFISfwpRo.wav", + "question": "Where is 'a vehicle engine running idle' found in its entirety within the audio's timeline?", + "choice_a": "[3.35, 5.7]", + "choice_b": "[5.72, 10.0]", + "choice_c": "[2.57, 3.35]", + "choice_d": "[4.09, 5.69]", + "answer_gt": "[3.35, 5.7]", + "answer_gt_word": "a vehicle engine running idle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11422 + }, + { + "path": "Yz1dShwq36TY.wav", + "question": "Identify the periods when 'engine chugging' fully resonates in the sound.", + "choice_a": "[0.01, 10.01]", + "choice_b": "[4.16, 5.59]", + "choice_c": "[1.25, 3.21]", + "choice_d": "[2.8, 4.5]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "engine chugging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11423 + }, + { + "path": "YzG0dfs6IAl0.wav", + "question": "At what timestamps does the complete 'humming of an oncoming train' occur?", + "choice_a": "[0.98, 1.6]", + "choice_b": "[4.66, 4.91]", + "choice_c": "[7.2, 7.64]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "humming of an oncoming train", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11424 + }, + { + "path": "YzX0OMIdZERU.wav", + "question": "At what points does 'laugh' exist completely in the timeline?", + "choice_a": "[2.5, 3.74]", + "choice_b": "[7.36, 9.53]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[4.34, 6.9]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "laugh", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11425 + }, + { + "path": "YzbeBikc3Qt8.wav", + "question": "What are the intervals in seconds of each 'low background speech' appearance?", + "choice_a": "[4.95, 6.82]", + "choice_b": "[7.4, 8.23]", + "choice_c": "[0.39, 4.96]", + "choice_d": "[4.95, 6.96]", + "answer_gt": "[0.39, 4.96]", + "answer_gt_word": "low background speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11426 + }, + { + "path": "YzsoRpeRkN34.wav", + "question": "At which time markers does 'an aircraft engine sharply running' fully unfold in the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.07, 7.05]", + "choice_c": "[0.11, 8.99]", + "choice_d": "[1.57, 9.14]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an aircraft engine sharply running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11427 + }, + { + "path": "Y-0A1_JR5f34.wav", + "question": "What are the full time segments when 'an aircraft engine sharply running' appears in this sound?", + "choice_a": "[0.29, 1.5]", + "choice_b": "[7.01, 8.24]", + "choice_c": "[1.51, 2.76]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an aircraft engine sharply running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11428 + }, + { + "path": "Y-2--Xzix2Vk.wav", + "question": "At what seconds do the full iterations of 'a cat meows' start and end?", + "choice_a": "[0.02, 0.87]", + "choice_b": "[5.36, 5.71]", + "choice_c": "[6.28, 6.51]", + "choice_d": "[1.18, 8.58]", + "answer_gt": "[0.02, 0.87]", + "answer_gt_word": "a cat meows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11429 + }, + { + "path": "Y-68677bCJWM.wav", + "question": "What are the boundaries in seconds for each 'talks' presence?", + "choice_a": "[0.0, 3.4]", + "choice_b": "[5.87, 8.48]", + "choice_c": "[5.48, 6.34]", + "choice_d": "[4.57, 5.3]", + "answer_gt": "[5.48, 6.34]", + "answer_gt_word": "talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11430 + }, + { + "path": "Y-68L7q5JlL4.wav", + "question": "At which seconds does 'wind against a microphone on a speeding car' fully register in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.04, 7.93]", + "choice_c": "[4.04, 8.66]", + "choice_d": "[5.64, 8.96]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind against a microphone on a speeding car", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11431 + }, + { + "path": "Y-A3K0_gtK4c.wav", + "question": "At what timestamps does the complete 'it blows its horn' occur?", + "choice_a": "[0.01, 8.0]", + "choice_b": "[5.09, 8.63]", + "choice_c": "[6.59, 9.45]", + "choice_d": "[8.35, 9.5]", + "answer_gt": "[0.01, 8.0]", + "answer_gt_word": "it blows its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11432 + }, + { + "path": "Y-ATrpUV4axE.wav", + "question": "At what timestamps does the complete 'the wind is blowing' occur?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.85, 9.4]", + "choice_c": "[4.33, 5.71]", + "choice_d": "[4.6, 6.69]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "the wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11433 + }, + { + "path": "Y-A_Gt60tfC8.wav", + "question": "What is the range in seconds for each full 'the wind is blowing' in the recording?", + "choice_a": "[0.01, 3.41]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.33, 3.94]", + "choice_d": "[7.46, 7.97]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "the wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11434 + }, + { + "path": "Y-BL3E3sLc78.wav", + "question": "Where does 'a whining noise in the background' appear from beginning to end in the sound?", + "choice_a": "[1.98, 2.93]", + "choice_b": "[1.2, 6.39]", + "choice_c": "[5.01, 7.75]", + "choice_d": "[5.92, 8.71]", + "answer_gt": "[1.98, 2.93]", + "answer_gt_word": "a whining noise in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11435 + }, + { + "path": "Y-I_Ej65k3Rg.wav", + "question": "What are the distinct time ranges of 'it barks' within the audio?", + "choice_a": "[0.3, 4.01]", + "choice_b": "[5.7, 6.84]", + "choice_c": "[7.53, 8.94]", + "choice_d": "[8.94, 9.66]", + "answer_gt": "[8.94, 9.66]", + "answer_gt_word": "it barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11436 + }, + { + "path": "Y-JaQwrOeCYs.wav", + "question": "What are the comprehensive timeframes of 'a steam whistle is blowing' in this clip?", + "choice_a": "[0.28, 10.0]", + "choice_b": "[5.24, 10.0]", + "choice_c": "[3.85, 9.12]", + "choice_d": "[7.36, 9.22]", + "answer_gt": "[0.28, 10.0]", + "answer_gt_word": "a steam whistle is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11437 + }, + { + "path": "Y-LKvu9q54f0.wav", + "question": "What time segments showcase 'a woman speaks' in its entirety?", + "choice_a": "[0.07, 5.02]", + "choice_b": "[7.78, 8.67]", + "choice_c": "[9.34, 9.99]", + "choice_d": "[6.1, 6.91]", + "answer_gt": "[9.34, 9.99]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11438 + }, + { + "path": "Y-NsC63dA01g.wav", + "question": "Identify the occurrences of 'a woman speaks' along with their time frames.", + "choice_a": "[9.34, 9.99]", + "choice_b": "[4.36, 5.15]", + "choice_c": "[6.37, 7.13]", + "choice_d": "[3.45, 3.98]", + "answer_gt": "[9.34, 9.99]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11439 + }, + { + "path": "Y-NsCNnryC9c.wav", + "question": "At what timestamps does the complete 'a woman speaks' occur?", + "choice_a": "[0.0, 0.89]", + "choice_b": "[1.91, 5.21]", + "choice_c": "[7.96, 8.18]", + "choice_d": "[9.34, 9.99]", + "answer_gt": "[9.34, 9.99]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11440 + }, + { + "path": "Y-S4jm_6mb3s.wav", + "question": "Identify the periods when 'a dog growling' fully resonates in the sound.", + "choice_a": "[1.99, 2.21]", + "choice_b": "[5.26, 6.3]", + "choice_c": "[7.99, 8.89]", + "choice_d": "[4.65, 9.28]", + "answer_gt": "[7.99, 8.89]", + "answer_gt_word": "a dog growling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11441 + }, + { + "path": "Y-S8UABOd2R8.wav", + "question": "How is 'revving and accelerating of an engine' distributed over the sound's duration?", + "choice_a": "[0.0, 7.04]", + "choice_b": "[0.48, 6.15]", + "choice_c": "[1.42, 6.59]", + "choice_d": "[4.61, 7.39]", + "answer_gt": "[0.0, 7.04]", + "answer_gt_word": "revving and accelerating of an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11442 + }, + { + "path": "Y-SuEOOhyRKs.wav", + "question": "Identify the periods when 'revving and accelerating of an engine' fully resonates in the sound.", + "choice_a": "[0.0, 7.04]", + "choice_b": "[1.76, 3.42]", + "choice_c": "[4.7, 6.03]", + "choice_d": "[7.23, 7.71]", + "answer_gt": "[0.0, 7.04]", + "answer_gt_word": "revving and accelerating of an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11443 + }, + { + "path": "Y-T1RU8oThuQ.wav", + "question": "What second marks correspond to the entire span of 'car speeding up'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.98, 9.11]", + "choice_d": "[5.95, 9.67]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "car speeding up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11444 + }, + { + "path": "Y-WFcBNcMBMQ.wav", + "question": "At which time markers does 'a man speaking' fully unfold in the sound?", + "choice_a": "[5.76, 7.85]", + "choice_b": "[2.57, 6.97]", + "choice_c": "[3.85, 7.52]", + "choice_d": "[4.47, 7.83]", + "answer_gt": "[5.76, 7.85]", + "answer_gt_word": "a man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11445 + }, + { + "path": "Y-XGsWplBxsk.wav", + "question": "What second marks correspond to the entire span of 'a car is driving by loudly'?", + "choice_a": "[0.0, 6.11]", + "choice_b": "[6.1, 7.66]", + "choice_c": "[0.46, 4.23]", + "choice_d": "[0.77, 7.81]", + "answer_gt": "[0.0, 6.11]", + "answer_gt_word": "a car is driving by loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11446 + }, + { + "path": "Y-_iFkUjP1Og.wav", + "question": "Where is 'a car is driving by loudly' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 6.11]", + "choice_b": "[5.75, 9.83]", + "choice_c": "[0.0, 5.92]", + "choice_d": "[7.67, 8.07]", + "answer_gt": "[0.0, 6.11]", + "answer_gt_word": "a car is driving by loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11447 + }, + { + "path": "Y-dpI6JsEyLs.wav", + "question": "At what points does 'water pouring' exist completely in the timeline?", + "choice_a": "[0.34, 1.94]", + "choice_b": "[2.76, 3.21]", + "choice_c": "[4.01, 5.53]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "water pouring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11448 + }, + { + "path": "Y-fl-wtFzC44.wav", + "question": "Capture the timeline where 'a woman groaning' is fully included in the sound.", + "choice_a": "[0.0, 5.79]", + "choice_b": "[6.68, 8.01]", + "choice_c": "[5.76, 6.72]", + "choice_d": "[5.76, 6.69]", + "answer_gt": "[5.76, 6.72]", + "answer_gt_word": "a woman groaning", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11449 + }, + { + "path": "Y-gT--kmfcqg.wav", + "question": "How many seconds does 'an engine running' span in each occurrence?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[3.78, 10.0]", + "choice_c": "[3.86, 6.68]", + "choice_d": "[4.99, 7.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an engine running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11450 + }, + { + "path": "Y-nJOtEC0tks.wav", + "question": "Define the time intervals that fully contain 'a medium motor vehicle engine is running' in this sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 1.01]", + "choice_c": "[5.54, 7.78]", + "choice_d": "[8.63, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a medium motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11451 + }, + { + "path": "Y-wDemwOuiuM.wav", + "question": "Capture the timeline where 'a medium motor vehicle engine is running' is fully included in the sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.01, 5.85]", + "choice_c": "[6.58, 8.13]", + "choice_d": "[1.68, 2.15]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a medium motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11452 + }, + { + "path": "Y-xieOaxGx8E.wav", + "question": "Where is 'a medium motor vehicle engine is running' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 1.29]", + "choice_b": "[2.41, 3.18]", + "choice_c": "[4.4, 5.4]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a medium motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11453 + }, + { + "path": "Y025yQh3lGLI.wav", + "question": "Identify the complete instances of 'a man talks briefly' within the sound timeline.", + "choice_a": "[0.0, 1.77]", + "choice_b": "[2.05, 2.31]", + "choice_c": "[1.15, 2.94]", + "choice_d": "[1.58, 3.23]", + "answer_gt": "[0.0, 1.77]", + "answer_gt_word": "a man talks briefly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11454 + }, + { + "path": "Y02yX3kyPZqU.wav", + "question": "In what time spans is 'speaking' completely audible in this recording?", + "choice_a": "[3.8, 9.99]", + "choice_b": "[2.16, 5.58]", + "choice_c": "[2.31, 6.72]", + "choice_d": "[4.52, 9.05]", + "answer_gt": "[3.8, 9.99]", + "answer_gt_word": "speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11455 + }, + { + "path": "Y05zALRKaib0.wav", + "question": "Determine the full durations of 'adult females are laughing' in the audio file.", + "choice_a": "[0.0, 0.64]", + "choice_b": "[1.31, 2.55]", + "choice_c": "[4.93, 10.0]", + "choice_d": "[2.7, 4.31]", + "answer_gt": "[2.7, 4.31]", + "answer_gt_word": "adult females are laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11456 + }, + { + "path": "Y0632OqvXrwg.wav", + "question": "Locate every segment where 'a man laughs' is completely audible.", + "choice_a": "[0.36, 1.1]", + "choice_b": "[1.32, 2.4]", + "choice_c": "[5.0, 5.24]", + "choice_d": "[2.79, 3.91]", + "answer_gt": "[0.36, 1.1]", + "answer_gt_word": "a man laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11457 + }, + { + "path": "Y079vcq6drhg.wav", + "question": "In what segments can 'engines running' be heard from start to finish?", + "choice_a": "[4.37, 10.0]", + "choice_b": "[1.24, 4.76]", + "choice_c": "[0.54, 6.25]", + "choice_d": "[1.51, 6.56]", + "answer_gt": "[4.37, 10.0]", + "answer_gt_word": "engines running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11458 + }, + { + "path": "Y09hPW_9Q8Rw.wav", + "question": "Enumerate the second-based occurrences of 'engines running' in full.", + "choice_a": "[0.32, 3.36]", + "choice_b": "[5.25, 9.6]", + "choice_c": "[4.37, 10.0]", + "choice_d": "[6.08, 6.74]", + "answer_gt": "[4.37, 10.0]", + "answer_gt_word": "engines running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11459 + }, + { + "path": "Y0Bdw87rcOdg.wav", + "question": "How long in seconds does 'a large vehicle engine is idling' last from start to end?", + "choice_a": "[0.01, 9.05]", + "choice_b": "[9.55, 10.0]", + "choice_c": "[2.54, 4.82]", + "choice_d": "[2.55, 5.64]", + "answer_gt": "[0.01, 9.05]", + "answer_gt_word": "a large vehicle engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11460 + }, + { + "path": "Y0BwgEjw2aCs.wav", + "question": "What are the boundaries in seconds for each 'a large vehicle engine is idling' presence?", + "choice_a": "[1.4, 3.95]", + "choice_b": "[0.01, 9.05]", + "choice_c": "[5.85, 6.9]", + "choice_d": "[2.4, 2.81]", + "answer_gt": "[0.01, 9.05]", + "answer_gt_word": "a large vehicle engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11461 + }, + { + "path": "Y0EVyUEVloBc.wav", + "question": "What are the specific start and stop times of 'a small train horn blows' in the sound?", + "choice_a": "[7.46, 10.02]", + "choice_b": "[0.23, 6.15]", + "choice_c": "[1.35, 8.78]", + "choice_d": "[1.83, 9.19]", + "answer_gt": "[7.46, 10.02]", + "answer_gt_word": "a small train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11462 + }, + { + "path": "Y0GRzxYNqvjI.wav", + "question": "Document the full appearances of 'honking' in terms of seconds.", + "choice_a": "[0.0, 8.28]", + "choice_b": "[9.53, 10.0]", + "choice_c": "[8.27, 9.07]", + "choice_d": "[6.23, 8.66]", + "answer_gt": "[8.27, 9.07]", + "answer_gt_word": "honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11463 + }, + { + "path": "Y0HDD9x7hsZ8.wav", + "question": "List the periods when 'high revving engine in front of crowd' is completely played in the sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.38, 3.42]", + "choice_c": "[1.38, 5.91]", + "choice_d": "[1.99, 7.12]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "high revving engine in front of crowd", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11464 + }, + { + "path": "Y0HMa_BIbTo0.wav", + "question": "What is the range in seconds for each full 'high revving engine in front of crowd' in the recording?", + "choice_a": "[0.0, 3.42]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.0, 3.4]", + "choice_d": "[4.72, 10.0]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "high revving engine in front of crowd", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11465 + }, + { + "path": "Y0HRItbIdDak.wav", + "question": "Identify the occurrences of 'small engine revving up and down' along with their time frames.", + "choice_a": "[2.5, 5.88]", + "choice_b": "[0.0, 2.48]", + "choice_c": "[5.96, 10.02]", + "choice_d": "[1.44, 4.62]", + "answer_gt": "[2.5, 5.88]", + "answer_gt_word": "small engine revving up and down", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11466 + }, + { + "path": "Y0IGsVL49I2M.wav", + "question": "Where does 'a dog barks' appear from beginning to end in the sound?", + "choice_a": "[0.11, 0.66]", + "choice_b": "[0.0, 0.18]", + "choice_c": "[0.78, 1.56]", + "choice_d": "[3.33, 10.0]", + "answer_gt": "[0.11, 0.66]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11467 + }, + { + "path": "Y0NvU5eiSlJw.wav", + "question": "Where do we hear 'wind whipping by outside' in full within this sound recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.19, 2.07]", + "choice_c": "[0.59, 6.39]", + "choice_d": "[1.49, 6.61]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind whipping by outside", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11468 + }, + { + "path": "Y0ZIy0COr3NM.wav", + "question": "How long does 'a woman laughs' last each time it appears in the audio?", + "choice_a": "[1.39, 4.03]", + "choice_b": "[4.62, 5.46]", + "choice_c": "[6.54, 8.52]", + "choice_d": "[4.13, 4.85]", + "answer_gt": "[6.54, 8.52]", + "answer_gt_word": "a woman laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11469 + }, + { + "path": "Y0_zLL39nfoM.wav", + "question": "List the specific seconds of 'a motor vehicle engine is running' appearances.", + "choice_a": "[0.98, 9.72]", + "choice_b": "[0.98, 9.72]", + "choice_c": "[3.11, 8.24]", + "choice_d": "[3.26, 9.14]", + "answer_gt": "[0.98, 9.72]", + "answer_gt_word": "a motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11470 + }, + { + "path": "Y0fPqbvlhkOc.wav", + "question": "At what time intervals does 'a dog pants' occur in the sound file?", + "choice_a": "[0.0, 8.14]", + "choice_b": "[1.13, 2.1]", + "choice_c": "[0.01, 0.74]", + "choice_d": "[0.95, 1.42]", + "answer_gt": "[0.0, 8.14]", + "answer_gt_word": "a dog pants", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11471 + }, + { + "path": "Y0g1DCi18pZA.wav", + "question": "In what segments can 'a young female speaks' be heard from start to finish?", + "choice_a": "[6.68, 7.08]", + "choice_b": "[7.72, 8.04]", + "choice_c": "[1.45, 5.08]", + "choice_d": "[2.53, 7.29]", + "answer_gt": "[6.68, 7.08]", + "answer_gt_word": "a young female speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11472 + }, + { + "path": "Y0gfxwS7Wla8.wav", + "question": "What are the specific start and stop times of 'vehicles engine accelerating' in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.75, 7.76]", + "choice_c": "[1.77, 4.54]", + "choice_d": "[3.15, 8.44]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "vehicles engine accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11473 + }, + { + "path": "Y0hKEM6Tq6zA.wav", + "question": "How is 'roaring' distributed over the sound's duration?", + "choice_a": "[1.79, 2.0]", + "choice_b": "[8.69, 8.83]", + "choice_c": "[2.83, 8.42]", + "choice_d": "[2.86, 5.02]", + "answer_gt": "[2.83, 8.42]", + "answer_gt_word": "roaring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11474 + }, + { + "path": "Y0jS_ViL5QmY.wav", + "question": "Which sections of the sound contain 'roaring' in their entirety?", + "choice_a": "[0.0, 0.68]", + "choice_b": "[1.38, 3.63]", + "choice_c": "[6.23, 9.2]", + "choice_d": "[2.83, 8.42]", + "answer_gt": "[2.83, 8.42]", + "answer_gt_word": "roaring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11475 + }, + { + "path": "Y0jTGXM7yEqs.wav", + "question": "What are the comprehensive timeframes of 'roaring' in this clip?", + "choice_a": "[0.08, 1.81]", + "choice_b": "[3.61, 4.46]", + "choice_c": "[2.83, 8.42]", + "choice_d": "[8.51, 8.95]", + "answer_gt": "[2.83, 8.42]", + "answer_gt_word": "roaring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11476 + }, + { + "path": "Y0lQ8ZEfBKlI.wav", + "question": "Capture the timeline where 'a person is starting up a motorcycle' is fully included in the sound.", + "choice_a": "[3.81, 10.0]", + "choice_b": "[2.22, 6.35]", + "choice_c": "[2.41, 7.17]", + "choice_d": "[3.07, 9.29]", + "answer_gt": "[3.81, 10.0]", + "answer_gt_word": "a person is starting up a motorcycle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11477 + }, + { + "path": "Y0nXHO9hRfO4.wav", + "question": "What time segments delineate the presence of 'barks'?", + "choice_a": "[0.34, 1.39]", + "choice_b": "[7.11, 8.46]", + "choice_c": "[5.18, 7.27]", + "choice_d": "[8.64, 9.47]", + "answer_gt": "[7.11, 8.46]", + "answer_gt_word": "barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11478 + }, + { + "path": "Y0pvQ0fpTmW8.wav", + "question": "Provide the start and end times of 'barks' within this audio.", + "choice_a": "[7.11, 8.46]", + "choice_b": "[1.46, 3.64]", + "choice_c": "[4.16, 4.73]", + "choice_d": "[5.76, 9.86]", + "answer_gt": "[7.11, 8.46]", + "answer_gt_word": "barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11479 + }, + { + "path": "Y0uRkdMtCuDM.wav", + "question": "What ranges in seconds capture 'squeaky sound plays' in full?", + "choice_a": "[4.84, 8.3]", + "choice_b": "[9.2, 10.0]", + "choice_c": "[2.73, 9.79]", + "choice_d": "[7.95, 8.97]", + "answer_gt": "[2.73, 9.79]", + "answer_gt_word": "squeaky sound plays", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11480 + }, + { + "path": "Y0wgFRL-xcd0.wav", + "question": "Define the time intervals that fully contain 'a dog barks' in this sound.", + "choice_a": "[8.83, 9.46]", + "choice_b": "[5.41, 6.13]", + "choice_c": "[7.01, 7.36]", + "choice_d": "[7.89, 9.79]", + "answer_gt": "[8.83, 9.46]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11481 + }, + { + "path": "Y0yETgW44MZU.wav", + "question": "At what timestamps does the complete 'more horn honking' occur?", + "choice_a": "[7.52, 10.0]", + "choice_b": "[0.01, 6.25]", + "choice_c": "[7.47, 10.0]", + "choice_d": "[6.42, 7.5]", + "answer_gt": "[7.52, 10.0]", + "answer_gt_word": "more horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11482 + }, + { + "path": "Y0zOfqWQuies.wav", + "question": "At what seconds do the full iterations of 'the man speaks' start and end?", + "choice_a": "[0.09, 0.76]", + "choice_b": "[2.26, 3.14]", + "choice_c": "[8.2, 9.31]", + "choice_d": "[9.24, 10.0]", + "answer_gt": "[8.2, 9.31]", + "answer_gt_word": "the man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11483 + }, + { + "path": "Y0zmN7Lvm_J0.wav", + "question": "What are the exact times when 'man speaking' is present in the clip?", + "choice_a": "[0.72, 2.06]", + "choice_b": "[8.79, 10.0]", + "choice_c": "[0.41, 7.19]", + "choice_d": "[1.5, 9.5]", + "answer_gt": "[0.72, 2.06]", + "answer_gt_word": "man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11484 + }, + { + "path": "Y11cctCD0ATQ.wav", + "question": "List the specific seconds of 'in the background a vehicle is running' appearances.", + "choice_a": "[0.0, 1.1]", + "choice_b": "[1.93, 8.96]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[2.89, 7.33]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "in the background a vehicle is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11485 + }, + { + "path": "Y11qcdPN80x8.wav", + "question": "What are the boundaries in seconds for each 'a cat meows' presence?", + "choice_a": "[1.28, 1.92]", + "choice_b": "[0.66, 1.04]", + "choice_c": "[1.42, 5.09]", + "choice_d": "[1.49, 7.98]", + "answer_gt": "[1.28, 1.92]", + "answer_gt_word": "a cat meows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11486 + }, + { + "path": "Y13QUBj_VlIc.wav", + "question": "Where is 'woman speaks' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 1.23]", + "choice_b": "[3.25, 4.25]", + "choice_c": "[8.5, 9.27]", + "choice_d": "[3.09, 4.44]", + "answer_gt": "[3.09, 4.44]", + "answer_gt_word": "woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11487 + }, + { + "path": "Y13WnlxzBU6k.wav", + "question": "How many seconds does 'woman speaks' span in each occurrence?", + "choice_a": "[0.14, 1.16]", + "choice_b": "[3.88, 10.0]", + "choice_c": "[1.06, 3.88]", + "choice_d": "[3.09, 4.44]", + "answer_gt": "[3.09, 4.44]", + "answer_gt_word": "woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11488 + }, + { + "path": "Y1Id9Ow6_v1M.wav", + "question": "Identify the complete instances of 'a person coughs' within the sound timeline.", + "choice_a": "[4.9, 5.38]", + "choice_b": "[6.04, 6.71]", + "choice_c": "[1.07, 7.04]", + "choice_d": "[3.04, 8.0]", + "answer_gt": "[4.9, 5.38]", + "answer_gt_word": "a person coughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11489 + }, + { + "path": "Y1NVfl7AnSRg.wav", + "question": "What are the comprehensive timeframes of 'a man is speaking' in this clip?", + "choice_a": "[1.28, 2.41]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.11, 3.21]", + "choice_d": "[2.78, 9.3]", + "answer_gt": "[1.28, 2.41]", + "answer_gt_word": "a man is speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11490 + }, + { + "path": "Y1QR8acH95Ho.wav", + "question": "What second marks correspond to the entire span of 'a man is speaking'?", + "choice_a": "[0.0, 3.83]", + "choice_b": "[5.28, 10.0]", + "choice_c": "[1.28, 2.41]", + "choice_d": "[8.33, 10.0]", + "answer_gt": "[1.28, 2.41]", + "answer_gt_word": "a man is speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11491 + }, + { + "path": "Y1R14_3kTmbs.wav", + "question": "Define the time intervals that fully contain 'a vehicle 's engine is starting' in this sound.", + "choice_a": "[0.89, 6.98]", + "choice_b": "[1.05, 3.93]", + "choice_c": "[2.28, 3.95]", + "choice_d": "[2.92, 5.09]", + "answer_gt": "[0.89, 6.98]", + "answer_gt_word": "a vehicle 's engine is starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11492 + }, + { + "path": "Y1RpnBTT-dxU.wav", + "question": "Identify the complete instances of 'a vehicle 's engine is starting' within the sound timeline.", + "choice_a": "[0.5, 2.65]", + "choice_b": "[3.23, 4.41]", + "choice_c": "[0.89, 6.98]", + "choice_d": "[0.86, 2.36]", + "answer_gt": "[0.89, 6.98]", + "answer_gt_word": "a vehicle 's engine is starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11493 + }, + { + "path": "Y1VQZAJgvzWI.wav", + "question": "Capture the timeline where 'some faint cracking noises' is fully included in the sound.", + "choice_a": "[0.0, 9.28]", + "choice_b": "[9.28, 10.0]", + "choice_c": "[0.55, 4.93]", + "choice_d": "[3.21, 8.94]", + "answer_gt": "[0.0, 9.28]", + "answer_gt_word": "some faint cracking noises", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11494 + }, + { + "path": "Y1VVr6_KQGwc.wav", + "question": "At what points does 'water runs' exist completely in the timeline?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.21, 6.16]", + "choice_c": "[8.55, 9.4]", + "choice_d": "[2.73, 9.09]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "water runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11495 + }, + { + "path": "Y1WzI9RX3CnE.wav", + "question": "Identify the complete instances of 'a train honks' within the sound timeline.", + "choice_a": "[1.47, 2.0]", + "choice_b": "[0.85, 5.7]", + "choice_c": "[3.94, 6.6]", + "choice_d": "[5.67, 8.36]", + "answer_gt": "[1.47, 2.0]", + "answer_gt_word": "a train honks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11496 + }, + { + "path": "Y1b7WUzSzU28.wav", + "question": "How long in seconds does 'water splashing' last from start to end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.0, 2.19]", + "choice_d": "[3.5, 4.09]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "water splashing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11497 + }, + { + "path": "Y1bAV_svnjUo.wav", + "question": "How are the complete segments of 'dogs are barking' timed in this audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[0.57, 9.99]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "dogs are barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11498 + }, + { + "path": "Y1dTDTUvU7yE.wav", + "question": "What are the comprehensive segments of 'a motor vehicle engine is running' in time?", + "choice_a": "[0.0, 9.58]", + "choice_b": "[3.66, 8.84]", + "choice_c": "[0.01, 6.15]", + "choice_d": "[6.51, 9.58]", + "answer_gt": "[0.0, 9.58]", + "answer_gt_word": "a motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11499 + }, + { + "path": "Y1dmz0KR31Xg.wav", + "question": "What are all the complete time segments of 'a motor vehicle engine is running' in this sound?", + "choice_a": "[0.0, 0.95]", + "choice_b": "[2.32, 3.14]", + "choice_c": "[4.66, 5.18]", + "choice_d": "[0.0, 9.58]", + "answer_gt": "[0.0, 9.58]", + "answer_gt_word": "a motor vehicle engine is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11500 + }, + { + "path": "Y1fwJTuGNNB0.wav", + "question": "At what timestamps does the complete 'a dog crying' occur?", + "choice_a": "[0.0, 7.85]", + "choice_b": "[0.0, 7.83]", + "choice_c": "[1.61, 8.39]", + "choice_d": "[6.13, 9.1]", + "answer_gt": "[0.0, 7.85]", + "answer_gt_word": "a dog crying", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11501 + }, + { + "path": "Y1iK9FP1ltuE.wav", + "question": "Determine the full durations of 'high-pitched engine starting' in the audio file.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.55, 7.89]", + "choice_d": "[7.28, 9.23]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "high-pitched engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11502 + }, + { + "path": "Y1iWFlLpixKU.wav", + "question": "Which sections of the sound contain 'high-pitched engine starting' in their entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.2, 9.8]", + "choice_c": "[2.2, 4.87]", + "choice_d": "[5.3, 7.65]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "high-pitched engine starting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11503 + }, + { + "path": "Y1kxnTHW60PQ.wav", + "question": "Where is 'a woman speaks' found in its entirety within the audio's timeline?", + "choice_a": "[3.39, 3.98]", + "choice_b": "[3.98, 4.62]", + "choice_c": "[1.99, 9.05]", + "choice_d": "[7.97, 9.54]", + "answer_gt": "[3.39, 3.98]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11504 + }, + { + "path": "Y1otuMUIZ6As.wav", + "question": "Identify the occurrences of 'vehicle engine noise' along with their time frames.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[4.72, 6.07]", + "choice_c": "[7.5, 7.93]", + "choice_d": "[8.83, 9.23]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "vehicle engine noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11505 + }, + { + "path": "Y1p8YDM6gG6Y.wav", + "question": "How many seconds does 'vehicle engine noise' span in each occurrence?", + "choice_a": "[0.0, 0.93]", + "choice_b": "[3.3, 4.04]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[5.62, 7.31]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "vehicle engine noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11506 + }, + { + "path": "Y1pDw9AoyZKk.wav", + "question": "What are the full time segments when 'a motor vehicle engine runs' appears in this sound?", + "choice_a": "[0.0, 9.81]", + "choice_b": "[4.45, 9.81]", + "choice_c": "[5.78, 9.81]", + "choice_d": "[1.01, 9.7]", + "answer_gt": "[0.0, 9.81]", + "answer_gt_word": "a motor vehicle engine runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11507 + }, + { + "path": "Y1pXuFVBomP8.wav", + "question": "Define the time intervals that fully contain 'a violin sounds' in this sound.", + "choice_a": "[0.12, 9.23]", + "choice_b": "[0.0, 5.24]", + "choice_c": "[6.06, 9.93]", + "choice_d": "[0.0, 5.24]", + "answer_gt": "[0.12, 9.23]", + "answer_gt_word": "a violin sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11508 + }, + { + "path": "Y1rPv2-OZLK4.wav", + "question": "Where is 'a violin sounds' consistently heard throughout the sound?", + "choice_a": "[0.0, 3.29]", + "choice_b": "[0.12, 9.23]", + "choice_c": "[1.98, 2.74]", + "choice_d": "[5.38, 6.38]", + "answer_gt": "[0.12, 9.23]", + "answer_gt_word": "a violin sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11509 + }, + { + "path": "Y1slvoNgzBLE.wav", + "question": "What ranges in seconds capture 'a subway train signal plays' in full?", + "choice_a": "[0.0, 2.55]", + "choice_b": "[6.41, 8.44]", + "choice_c": "[3.87, 5.03]", + "choice_d": "[9.59, 10.0]", + "answer_gt": "[0.0, 2.55]", + "answer_gt_word": "a subway train signal plays", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11510 + }, + { + "path": "Y1z08ytHRqfA.wav", + "question": "What are the full time segments when 'male' appears in this sound?", + "choice_a": "[0.0, 2.65]", + "choice_b": "[6.74, 10.0]", + "choice_c": "[2.97, 7.58]", + "choice_d": "[4.53, 8.3]", + "answer_gt": "[0.0, 2.65]", + "answer_gt_word": "male", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11511 + }, + { + "path": "Y24pz7PVwcw0.wav", + "question": "What are the precise second intervals of 'a crowd is talking' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.79, 8.49]", + "choice_c": "[2.72, 6.81]", + "choice_d": "[6.67, 8.57]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a crowd is talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11512 + }, + { + "path": "Y26dNsDuIt9Q.wav", + "question": "Document the full appearances of 'honks its horn' in terms of seconds.", + "choice_a": "[9.23, 10.0]", + "choice_b": "[3.77, 7.46]", + "choice_c": "[5.19, 8.9]", + "choice_d": "[7.0, 9.1]", + "answer_gt": "[9.23, 10.0]", + "answer_gt_word": "honks its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11513 + }, + { + "path": "Y28XfXlC7zkw.wav", + "question": "At what seconds do the full iterations of 'an engine lowly hums' start and end?", + "choice_a": "[0.0, 10.01]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[0.25, 5.46]", + "choice_d": "[4.36, 6.98]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an engine lowly hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11514 + }, + { + "path": "Y2AMWVZMp_Xs.wav", + "question": "What ranges in seconds capture 'an engine lowly hums' in full?", + "choice_a": "[5.44, 6.01]", + "choice_b": "[9.74, 10.0]", + "choice_c": "[4.33, 4.94]", + "choice_d": "[0.0, 10.01]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an engine lowly hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11515 + }, + { + "path": "Y2Awp9Nvsb2A.wav", + "question": "At what time intervals does 'an engine lowly hums' occur in the sound file?", + "choice_a": "[0.24, 2.21]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[5.47, 5.84]", + "choice_d": "[7.6, 8.15]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an engine lowly hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11516 + }, + { + "path": "Y2BZwHNWM0_w.wav", + "question": "Determine the full durations of 'a car revs' in the audio file.", + "choice_a": "[0.0, 1.53]", + "choice_b": "[4.59, 5.18]", + "choice_c": "[5.85, 9.74]", + "choice_d": "[3.44, 4.54]", + "answer_gt": "[0.0, 1.53]", + "answer_gt_word": "a car revs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11517 + }, + { + "path": "Y2E8eCKwj2M4.wav", + "question": "At what time points is 'a dog barks' fully present in the recording?", + "choice_a": "[1.45, 3.56]", + "choice_b": "[0.06, 0.88]", + "choice_c": "[3.63, 7.63]", + "choice_d": "[8.13, 10.02]", + "answer_gt": "[1.45, 3.56]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11518 + }, + { + "path": "Y2GBJOdvh9XM.wav", + "question": "How are the complete segments of 'dogs bark' timed in this audio?", + "choice_a": "[0.0, 1.2]", + "choice_b": "[2.06, 2.46]", + "choice_c": "[3.87, 4.62]", + "choice_d": "[0.48, 10.0]", + "answer_gt": "[0.48, 10.0]", + "answer_gt_word": "dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11519 + }, + { + "path": "Y2IQF6utML0I.wav", + "question": "At what time points is 'a loud horn' fully present in the recording?", + "choice_a": "[0.85, 1.78]", + "choice_b": "[0.79, 5.8]", + "choice_c": "[0.8, 6.88]", + "choice_d": "[4.38, 7.6]", + "answer_gt": "[0.85, 1.78]", + "answer_gt_word": "a loud horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11520 + }, + { + "path": "Y2KR0C5ysO8o.wav", + "question": "Locate every segment where 'an engine revving continuously' is completely audible.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.22, 2.5]", + "choice_c": "[0.37, 4.71]", + "choice_d": "[1.62, 4.96]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine revving continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11521 + }, + { + "path": "Y2N_9HjzCW40.wav", + "question": "At which time markers does 'an engine revving' fully unfold in the sound?", + "choice_a": "[0.6, 5.42]", + "choice_b": "[7.55, 8.53]", + "choice_c": "[0.55, 8.41]", + "choice_d": "[3.42, 8.5]", + "answer_gt": "[0.6, 5.42]", + "answer_gt_word": "an engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11522 + }, + { + "path": "Y2PzTYy5GfqI.wav", + "question": "Document the full appearances of 'an engine revving' in terms of seconds.", + "choice_a": "[0.17, 5.29]", + "choice_b": "[0.6, 5.42]", + "choice_c": "[8.94, 10.0]", + "choice_d": "[5.38, 6.27]", + "answer_gt": "[0.6, 5.42]", + "answer_gt_word": "an engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11523 + }, + { + "path": "Y2QsWNp64ERU.wav", + "question": "What are the precise second intervals of 'train horn beeping' in this clip?", + "choice_a": "[0.0, 5.26]", + "choice_b": "[5.57, 10.0]", + "choice_c": "[2.81, 5.98]", + "choice_d": "[3.15, 6.63]", + "answer_gt": "[0.0, 5.26]", + "answer_gt_word": "train horn beeping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11524 + }, + { + "path": "Y2SELYAzRvN0.wav", + "question": "Where is 'a motor is idling in the background' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[3.78, 7.7]", + "choice_c": "[4.23, 8.66]", + "choice_d": "[5.58, 9.83]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a motor is idling in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11525 + }, + { + "path": "Y2VJSUZg-8Mk.wav", + "question": "What time segments delineate the presence of 'a vehicle accelerates'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[9.43, 10.0]", + "choice_c": "[4.86, 6.64]", + "choice_d": "[5.66, 6.99]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a vehicle accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11526 + }, + { + "path": "Y2Xi_JpZkLs0.wav", + "question": "In what time spans is 'a vehicle accelerates' completely audible in this recording?", + "choice_a": "[0.0, 0.83]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[3.78, 6.71]", + "choice_d": "[9.11, 9.54]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a vehicle accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11527 + }, + { + "path": "Y2YzvStRrnkQ.wav", + "question": "Capture the timeline where 'a motorcycle revs its engine' is fully included in the sound.", + "choice_a": "[0.0, 8.29]", + "choice_b": "[2.37, 3.94]", + "choice_c": "[5.7, 6.5]", + "choice_d": "[8.15, 10.0]", + "answer_gt": "[0.0, 8.29]", + "answer_gt_word": "a motorcycle revs its engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11528 + }, + { + "path": "Y2eG8jc_R46E.wav", + "question": "At what timestamps does the complete 'an engine is idling' occur?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.86, 4.54]", + "choice_c": "[2.97, 6.82]", + "choice_d": "[4.54, 9.17]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11529 + }, + { + "path": "Y2er6dDbTYcM.wav", + "question": "What are the full time segments when 'footsteps on wood outside' appears in this sound?", + "choice_a": "[0.0, 4.7]", + "choice_b": "[4.4, 9.17]", + "choice_c": "[9.17, 10.0]", + "choice_d": "[0.99, 9.72]", + "answer_gt": "[0.0, 4.7]", + "answer_gt_word": "footsteps on wood outside", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11530 + }, + { + "path": "Y2gMs40vtC6w.wav", + "question": "Locate every segment where 'wind blowing' is completely audible.", + "choice_a": "[0.0, 9.6]", + "choice_b": "[4.27, 6.33]", + "choice_c": "[6.87, 7.91]", + "choice_d": "[7.91, 9.36]", + "answer_gt": "[0.0, 9.6]", + "answer_gt_word": "wind blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11531 + }, + { + "path": "Y2h_F8niDeKA.wav", + "question": "In what segments can 'people are talking' be heard from start to finish?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.73, 6.04]", + "choice_c": "[7.15, 7.42]", + "choice_d": "[8.19, 8.46]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11532 + }, + { + "path": "Y2swvx975qSY.wav", + "question": "What ranges in seconds capture 'a horse is running' in full?", + "choice_a": "[0.0, 6.15]", + "choice_b": "[4.83, 6.68]", + "choice_c": "[6.43, 6.7]", + "choice_d": "[5.35, 6.52]", + "answer_gt": "[0.0, 6.15]", + "answer_gt_word": "a horse is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11533 + }, + { + "path": "Y2uOCHpAzwXI.wav", + "question": "How long does 'a dog howling like a wolf' last each time it appears in the audio?", + "choice_a": "[0.83, 3.4]", + "choice_b": "[3.5, 10.0]", + "choice_c": "[5.09, 6.88]", + "choice_d": "[5.58, 7.03]", + "answer_gt": "[0.83, 3.4]", + "answer_gt_word": "a dog howling like a wolf", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11534 + }, + { + "path": "Y2ugsMuS9EXE.wav", + "question": "What is the range in seconds for each full 'speech' in the recording?", + "choice_a": "[0.65, 1.22]", + "choice_b": "[9.41, 9.86]", + "choice_c": "[4.14, 5.85]", + "choice_d": "[7.57, 7.94]", + "answer_gt": "[9.41, 9.86]", + "answer_gt_word": "speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11535 + }, + { + "path": "Y2vBbpFbRHmk.wav", + "question": "At what seconds do the full iterations of 'cats are meowing' start and end?", + "choice_a": "[3.72, 5.11]", + "choice_b": "[0.09, 0.65]", + "choice_c": "[0.78, 6.21]", + "choice_d": "[3.04, 6.72]", + "answer_gt": "[3.72, 5.11]", + "answer_gt_word": "cats are meowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11536 + }, + { + "path": "Y2wNAe_DNXI8.wav", + "question": "What are the comprehensive segments of 'train toots' in time?", + "choice_a": "[1.71, 2.05]", + "choice_b": "[2.09, 3.26]", + "choice_c": "[3.98, 5.82]", + "choice_d": "[1.55, 6.27]", + "answer_gt": "[1.71, 2.05]", + "answer_gt_word": "train toots", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11537 + }, + { + "path": "Y2wz_XyJ4Zxk.wav", + "question": "What time segments delineate the presence of 'train toots'?", + "choice_a": "[0.75, 3.0]", + "choice_b": "[4.88, 6.07]", + "choice_c": "[1.71, 2.05]", + "choice_d": "[3.72, 4.52]", + "answer_gt": "[1.71, 2.05]", + "answer_gt_word": "train toots", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11538 + }, + { + "path": "Y2xg_8rf9J50.wav", + "question": "How many seconds does each full presence of 'a dog barking' cover?", + "choice_a": "[1.21, 1.94]", + "choice_b": "[8.28, 8.54]", + "choice_c": "[9.46, 10.0]", + "choice_d": "[5.13, 10.02]", + "answer_gt": "[5.13, 10.02]", + "answer_gt_word": "a dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11539 + }, + { + "path": "Y3-9kjeKeew8.wav", + "question": "Document the full appearances of 'a dog barking' in terms of seconds.", + "choice_a": "[0.0, 0.32]", + "choice_b": "[5.13, 10.02]", + "choice_c": "[3.37, 4.38]", + "choice_d": "[0.77, 1.52]", + "answer_gt": "[5.13, 10.02]", + "answer_gt_word": "a dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11540 + }, + { + "path": "Y30D1aU785Gg.wav", + "question": "Where does 'a dog barking' appear from beginning to end in the sound?", + "choice_a": "[0.11, 2.24]", + "choice_b": "[5.13, 10.02]", + "choice_c": "[5.5, 6.21]", + "choice_d": "[2.39, 3.86]", + "answer_gt": "[5.13, 10.02]", + "answer_gt_word": "a dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11541 + }, + { + "path": "Y30Sm6NPl_rI.wav", + "question": "Provide the start and end times of 'multiple dogs bark' within this audio.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.58, 4.68]", + "choice_d": "[4.5, 4.9]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "multiple dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11542 + }, + { + "path": "Y31oUjXfb1Gk.wav", + "question": "Provide the start and end times of 'a small dog barking' within this audio.", + "choice_a": "[0.0, 3.83]", + "choice_b": "[4.56, 5.84]", + "choice_c": "[7.51, 8.95]", + "choice_d": "[0.52, 0.81]", + "answer_gt": "[0.52, 0.81]", + "answer_gt_word": "a small dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11543 + }, + { + "path": "Y32T8xTpvTFI.wav", + "question": "What is the range in seconds for each full 'an animal bleats' in the recording?", + "choice_a": "[0.0, 0.71]", + "choice_b": "[1.93, 6.82]", + "choice_c": "[3.19, 3.81]", + "choice_d": "[9.53, 10.0]", + "answer_gt": "[3.19, 3.81]", + "answer_gt_word": "an animal bleats", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11544 + }, + { + "path": "Y32vu_dTOVZM.wav", + "question": "At what time intervals does 'a woman speaks' occur in the sound file?", + "choice_a": "[2.59, 2.93]", + "choice_b": "[4.0, 5.05]", + "choice_c": "[5.39, 6.47]", + "choice_d": "[3.73, 5.54]", + "answer_gt": "[5.39, 6.47]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11545 + }, + { + "path": "Y350OCezayrk.wav", + "question": "Enumerate the second-based occurrences of 'a motor vehicle engine starter grinds' in full.", + "choice_a": "[0.0, 4.81]", + "choice_b": "[4.87, 10.0]", + "choice_c": "[3.31, 5.0]", + "choice_d": "[0.42, 2.83]", + "answer_gt": "[0.0, 4.81]", + "answer_gt_word": "a motor vehicle engine starter grinds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11546 + }, + { + "path": "Y35GZhbIN5lg.wav", + "question": "At which time markers does 'frantic speech' fully unfold in the sound?", + "choice_a": "[0.0, 5.68]", + "choice_b": "[2.96, 7.11]", + "choice_c": "[3.43, 5.08]", + "choice_d": "[4.82, 8.09]", + "answer_gt": "[0.0, 5.68]", + "answer_gt_word": "frantic speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11547 + }, + { + "path": "Y35t3MVCE160.wav", + "question": "What are the precise second intervals of 'police car siren' in this clip?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.39, 6.07]", + "choice_d": "[4.46, 8.22]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "police car siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11548 + }, + { + "path": "Y39EiNqi8feA.wav", + "question": "Where does 'dishes rattle' appear from beginning to end in the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[6.63, 7.48]", + "choice_c": "[0.55, 3.11]", + "choice_d": "[2.56, 8.37]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "dishes rattle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11549 + }, + { + "path": "Y3C9e_UT9ASs.wav", + "question": "What are the intervals in seconds of each 'a woman speaks' appearance?", + "choice_a": "[0.0, 3.97]", + "choice_b": "[4.38, 10.0]", + "choice_c": "[6.44, 8.46]", + "choice_d": "[8.94, 10.02]", + "answer_gt": "[0.0, 3.97]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11550 + }, + { + "path": "Y3DMi_CmlYJ4.wav", + "question": "How many seconds does each full presence of 'a woman speaks' cover?", + "choice_a": "[0.0, 3.45]", + "choice_b": "[4.31, 8.37]", + "choice_c": "[0.0, 3.97]", + "choice_d": "[0.22, 1.16]", + "answer_gt": "[0.0, 3.97]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11551 + }, + { + "path": "Y3JVI9QiahhE.wav", + "question": "How is 'a woman speaks' distributed over the sound's duration?", + "choice_a": "[0.06, 1.43]", + "choice_b": "[1.77, 10.0]", + "choice_c": "[0.0, 2.27]", + "choice_d": "[0.0, 3.97]", + "answer_gt": "[0.0, 3.97]", + "answer_gt_word": "a woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11552 + }, + { + "path": "Y3Lcxs4pPY9I.wav", + "question": "How is 'wind blowing hard' distributed over the sound's duration?", + "choice_a": "[0.01, 8.05]", + "choice_b": "[4.31, 8.06]", + "choice_c": "[2.83, 8.11]", + "choice_d": "[5.56, 8.93]", + "answer_gt": "[0.01, 8.05]", + "answer_gt_word": "wind blowing hard", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11553 + }, + { + "path": "Y3NPidVvhTnI.wav", + "question": "What times mark the complete cycles of 'heavy wind' in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.26, 2.94]", + "choice_d": "[2.67, 6.52]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy wind", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11554 + }, + { + "path": "Y3RPKGqAHRiU.wav", + "question": "Provide the start and end times of 'heavy wind' within this audio.", + "choice_a": "[0.41, 1.32]", + "choice_b": "[1.58, 2.42]", + "choice_c": "[2.54, 2.99]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy wind", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11555 + }, + { + "path": "Y3RZySuDzocQ.wav", + "question": "What times mark the complete cycles of 'a vehicle approaches very quickly' in the audio?", + "choice_a": "[0.0, 8.93]", + "choice_b": "[1.44, 8.26]", + "choice_c": "[2.78, 8.79]", + "choice_d": "[4.94, 9.35]", + "answer_gt": "[0.0, 8.93]", + "answer_gt_word": "a vehicle approaches very quickly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11556 + }, + { + "path": "Y3U1ZravTj_Y.wav", + "question": "At what time intervals does 'a vehicle is revving and accelerating' occur in the sound file?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.96, 5.52]", + "choice_c": "[1.7, 8.82]", + "choice_d": "[3.45, 9.03]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a vehicle is revving and accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11557 + }, + { + "path": "Y3UjD8ncozAA.wav", + "question": "Locate every segment where 'a quiet bang' is completely audible.", + "choice_a": "[1.82, 2.4]", + "choice_b": "[7.23, 8.29]", + "choice_c": "[4.87, 6.74]", + "choice_d": "[7.88, 10.0]", + "answer_gt": "[1.82, 2.4]", + "answer_gt_word": "a quiet bang", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11558 + }, + { + "path": "Y3a3cnDXnb1k.wav", + "question": "How long does 'a bunch of dogs barking in the background' last each time it appears in the audio?", + "choice_a": "[0.0, 4.38]", + "choice_b": "[5.34, 5.81]", + "choice_c": "[6.55, 6.92]", + "choice_d": "[0.0, 9.91]", + "answer_gt": "[0.0, 9.91]", + "answer_gt_word": "a bunch of dogs barking in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11559 + }, + { + "path": "Y3dFdo47pK1Q.wav", + "question": "Capture the timeline where 'a small dog barking' is fully included in the sound.", + "choice_a": "[0.27, 4.88]", + "choice_b": "[3.12, 5.74]", + "choice_c": "[4.54, 5.91]", + "choice_d": "[5.56, 6.27]", + "answer_gt": "[0.27, 4.88]", + "answer_gt_word": "a small dog barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11560 + }, + { + "path": "Y3ek_TD_ZAWA.wav", + "question": "What are the comprehensive timeframes of 'a car engine roars' in this clip?", + "choice_a": "[0.0, 9.96]", + "choice_b": "[0.11, 2.26]", + "choice_c": "[1.62, 2.76]", + "choice_d": "[2.08, 7.86]", + "answer_gt": "[0.0, 9.96]", + "answer_gt_word": "a car engine roars", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11561 + }, + { + "path": "Y3n4G4MbYRFE.wav", + "question": "Which sections of the sound contain 'engine chugging' in their entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.34, 4.24]", + "choice_c": "[7.21, 10.0]", + "choice_d": "[4.28, 8.09]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "engine chugging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11562 + }, + { + "path": "Y3ndid3jni7M.wav", + "question": "How long in seconds does 'a horn honking' last from start to end?", + "choice_a": "[0.83, 3.57]", + "choice_b": "[8.13, 10.0]", + "choice_c": "[3.55, 7.98]", + "choice_d": "[8.72, 10.0]", + "answer_gt": "[3.55, 7.98]", + "answer_gt_word": "a horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11563 + }, + { + "path": "Y3ngXHZYzxww.wav", + "question": "How long in seconds does 'a horn honking' last from start to end?", + "choice_a": "[0.81, 2.57]", + "choice_b": "[3.55, 7.98]", + "choice_c": "[5.67, 7.31]", + "choice_d": "[9.25, 9.88]", + "answer_gt": "[3.55, 7.98]", + "answer_gt_word": "a horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11564 + }, + { + "path": "Y3obLPXCKYpM.wav", + "question": "What second marks correspond to the entire span of 'multiple car engines roar'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.48, 7.88]", + "choice_c": "[5.82, 8.32]", + "choice_d": "[6.27, 8.5]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "multiple car engines roar", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11565 + }, + { + "path": "Y3qnpOqvnKTo.wav", + "question": "How are the complete segments of 'rumbling of a train passes by' timed in this audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[5.39, 7.89]", + "choice_c": "[0.42, 0.89]", + "choice_d": "[3.14, 5.39]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rumbling of a train passes by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11566 + }, + { + "path": "Y3rAJOlCARSg.wav", + "question": "Where do we hear 'a dog barks' in full within this sound recording?", + "choice_a": "[0.75, 1.17]", + "choice_b": "[6.37, 8.08]", + "choice_c": "[9.28, 10.0]", + "choice_d": "[1.54, 2.06]", + "answer_gt": "[0.75, 1.17]", + "answer_gt_word": "a dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11567 + }, + { + "path": "Y3rUTwDYoPDI.wav", + "question": "In what time spans is 'nature noises' completely audible in this recording?", + "choice_a": "[0.0, 6.11]", + "choice_b": "[6.1, 8.67]", + "choice_c": "[1.5, 7.12]", + "choice_d": "[6.2, 8.64]", + "answer_gt": "[0.0, 6.11]", + "answer_gt_word": "nature noises", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11568 + }, + { + "path": "Y3uOXUGT0kxY.wav", + "question": "What is the range in seconds for each full 'a group of people are speaking' in the recording?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.53, 1.32]", + "choice_c": "[2.18, 3.04]", + "choice_d": "[5.88, 6.25]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a group of people are speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11569 + }, + { + "path": "Y3uSEbfnfPUw.wav", + "question": "How many seconds does each full presence of 'panting' cover?", + "choice_a": "[2.91, 3.34]", + "choice_b": "[4.48, 4.93]", + "choice_c": "[5.14, 7.19]", + "choice_d": "[8.2, 8.81]", + "answer_gt": "[5.14, 7.19]", + "answer_gt_word": "panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11570 + }, + { + "path": "Y3wVoCVgkqDE.wav", + "question": "Provide the start and end times of 'panting' within this audio.", + "choice_a": "[0.73, 5.17]", + "choice_b": "[7.5, 8.0]", + "choice_c": "[5.14, 7.19]", + "choice_d": "[7.69, 8.55]", + "answer_gt": "[5.14, 7.19]", + "answer_gt_word": "panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11571 + }, + { + "path": "Y3xLintYp7PY.wav", + "question": "What second marks correspond to the entire span of 'a cat meows'?", + "choice_a": "[3.47, 3.88]", + "choice_b": "[3.96, 4.78]", + "choice_c": "[5.49, 6.37]", + "choice_d": "[6.87, 9.03]", + "answer_gt": "[3.47, 3.88]", + "answer_gt_word": "a cat meows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11572 + }, + { + "path": "Y3xepWaHuMN0.wav", + "question": "What is the range in seconds for each full 'a train blows its horn' in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.34, 6.09]", + "choice_c": "[0.86, 6.99]", + "choice_d": "[2.11, 8.45]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a train blows its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11573 + }, + { + "path": "Y40cuHrYfaqA.wav", + "question": "Define the time intervals that fully contain 'a train blows its horn' in this sound.", + "choice_a": "[0.04, 0.28]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[3.57, 3.8]", + "choice_d": "[0.04, 0.29]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a train blows its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11574 + }, + { + "path": "Y41LCAS35xP8.wav", + "question": "Identify the periods when 'engine idling noise' fully resonates in the sound.", + "choice_a": "[0.0, 9.02]", + "choice_b": "[2.08, 6.21]", + "choice_c": "[2.21, 7.47]", + "choice_d": "[3.18, 8.77]", + "answer_gt": "[0.0, 9.02]", + "answer_gt_word": "engine idling noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11575 + }, + { + "path": "Y41_uGt-6cHI.wav", + "question": "Determine the full durations of 'a dog is barking' in the audio file.", + "choice_a": "[0.31, 0.61]", + "choice_b": "[1.06, 1.42]", + "choice_c": "[2.67, 3.08]", + "choice_d": "[4.19, 4.71]", + "answer_gt": "[0.31, 0.61]", + "answer_gt_word": "a dog is barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11576 + }, + { + "path": "Y4GFeZgTcgRg.wav", + "question": "Enumerate the second-based occurrences of 'a dog is barking' in full.", + "choice_a": "[0.31, 0.61]", + "choice_b": "[3.54, 4.11]", + "choice_c": "[5.97, 6.66]", + "choice_d": "[8.28, 9.07]", + "answer_gt": "[0.31, 0.61]", + "answer_gt_word": "a dog is barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11577 + }, + { + "path": "Y4GalkerBVs8.wav", + "question": "Define the time intervals that fully contain 'a dog whimpers' in this sound.", + "choice_a": "[0.73, 9.6]", + "choice_b": "[0.71, 9.58]", + "choice_c": "[4.62, 7.72]", + "choice_d": "[2.85, 6.27]", + "answer_gt": "[0.73, 9.6]", + "answer_gt_word": "a dog whimpers", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11578 + }, + { + "path": "Y4GfN0gZQvXs.wav", + "question": "How are the complete segments of 'someone talks' timed in this audio?", + "choice_a": "[1.72, 4.25]", + "choice_b": "[3.75, 4.23]", + "choice_c": "[5.33, 10.0]", + "choice_d": "[2.68, 7.83]", + "answer_gt": "[1.72, 4.25]", + "answer_gt_word": "someone talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11579 + }, + { + "path": "Y4GzG-sDAwmI.wav", + "question": "In what segments can 'someone talks' be heard from start to finish?", + "choice_a": "[0.0, 0.48]", + "choice_b": "[2.17, 4.47]", + "choice_c": "[1.72, 4.25]", + "choice_d": "[2.17, 4.44]", + "answer_gt": "[1.72, 4.25]", + "answer_gt_word": "someone talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11580 + }, + { + "path": "Y4NhrtAEUsnc.wav", + "question": "What are the intervals in seconds of each 'helicopter blades spinning' appearance?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.83, 6.62]", + "choice_c": "[2.04, 8.18]", + "choice_d": "[3.54, 8.96]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "helicopter blades spinning", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11581 + }, + { + "path": "Y4OIHJwcKSWI.wav", + "question": "In what time spans is 'honk their horns' completely audible in this recording?", + "choice_a": "[2.65, 3.74]", + "choice_b": "[0.9, 6.33]", + "choice_c": "[1.92, 9.42]", + "choice_d": "[2.61, 9.66]", + "answer_gt": "[2.65, 3.74]", + "answer_gt_word": "honk their horns", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11582 + }, + { + "path": "Y4P7p_EIdIeY.wav", + "question": "Identify the periods when 'birds are singing in the background' fully resonates in the sound.", + "choice_a": "[0.0, 1.09]", + "choice_b": "[2.64, 3.55]", + "choice_c": "[4.01, 5.13]", + "choice_d": "[5.85, 8.53]", + "answer_gt": "[5.85, 8.53]", + "answer_gt_word": "birds are singing in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11583 + }, + { + "path": "Y4SGVC_RNijM.wav", + "question": "In what segments can 'a motor vehicle engine is idling' be heard from start to finish?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[6.48, 10.0]", + "choice_c": "[3.64, 5.47]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a motor vehicle engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11584 + }, + { + "path": "Y4SbXNmgnyio.wav", + "question": "How many seconds does each full presence of 'an engine revs fast' cover?", + "choice_a": "[0.65, 10.02]", + "choice_b": "[0.46, 3.06]", + "choice_c": "[2.38, 4.37]", + "choice_d": "[2.6, 9.9]", + "answer_gt": "[0.65, 10.02]", + "answer_gt_word": "an engine revs fast", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11585 + }, + { + "path": "Y4V3TFiHFqDc.wav", + "question": "How long in seconds does 'people are talking' last from start to end?", + "choice_a": "[0.8, 1.87]", + "choice_b": "[3.19, 8.36]", + "choice_c": "[9.5, 9.93]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11586 + }, + { + "path": "Y4Zcfo6EtDww.wav", + "question": "Document the full appearances of 'cat meowing' in terms of seconds.", + "choice_a": "[3.02, 4.88]", + "choice_b": "[0.24, 0.66]", + "choice_c": "[2.54, 2.83]", + "choice_d": "[5.67, 8.36]", + "answer_gt": "[3.02, 4.88]", + "answer_gt_word": "cat meowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11587 + }, + { + "path": "Y4_bKiFo_znw.wav", + "question": "Determine the full durations of 'cat meowing' in the audio file.", + "choice_a": "[3.02, 4.88]", + "choice_b": "[8.68, 9.32]", + "choice_c": "[6.05, 6.38]", + "choice_d": "[9.62, 10.0]", + "answer_gt": "[3.02, 4.88]", + "answer_gt_word": "cat meowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11588 + }, + { + "path": "Y4m5RneU2kHM.wav", + "question": "What timestamps mark the full occurrences of 'people are talking'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.08, 2.04]", + "choice_c": "[1.95, 5.91]", + "choice_d": "[2.01, 6.7]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people are talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11589 + }, + { + "path": "Y4msTzxbPaSc.wav", + "question": "List the periods when 'honks its horn' is completely played in the sound.", + "choice_a": "[2.74, 3.06]", + "choice_b": "[2.91, 7.35]", + "choice_c": "[3.64, 8.67]", + "choice_d": "[5.64, 9.32]", + "answer_gt": "[2.74, 3.06]", + "answer_gt_word": "honks its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11590 + }, + { + "path": "Y4ukueAa9oTk.wav", + "question": "At what seconds do the full iterations of 'growls' start and end?", + "choice_a": "[0.0, 1.41]", + "choice_b": "[2.57, 3.39]", + "choice_c": "[5.39, 6.75]", + "choice_d": "[7.5, 9.1]", + "answer_gt": "[7.5, 9.1]", + "answer_gt_word": "growls", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11591 + }, + { + "path": "Y4wjKtAs9aMQ.wav", + "question": "At what points does 'an engine is idling' exist completely in the timeline?", + "choice_a": "[0.0, 9.52]", + "choice_b": "[0.31, 3.6]", + "choice_c": "[1.34, 4.93]", + "choice_d": "[2.4, 8.84]", + "answer_gt": "[0.0, 9.52]", + "answer_gt_word": "an engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11592 + }, + { + "path": "Y4xrL4TSgHwU.wav", + "question": "What are the exact times when 'loud banging' is present in the clip?", + "choice_a": "[0.0, 5.25]", + "choice_b": "[5.2, 10.0]", + "choice_c": "[1.23, 6.91]", + "choice_d": "[1.44, 7.54]", + "answer_gt": "[0.0, 5.25]", + "answer_gt_word": "loud banging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11593 + }, + { + "path": "Y4xrL4TSgHwU.wav", + "question": "How is 'running idle' distributed over the sound's duration?", + "choice_a": "[0.01, 10.01]", + "choice_b": "[1.77, 8.47]", + "choice_c": "[3.36, 9.23]", + "choice_d": "[3.39, 9.64]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "running idle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11594 + }, + { + "path": "Y4yv--RtWtHg.wav", + "question": "Identify the periods when 'dog bark' fully resonates in the sound.", + "choice_a": "[0.0, 1.06]", + "choice_b": "[1.4, 1.83]", + "choice_c": "[5.4, 8.03]", + "choice_d": "[8.93, 10.0]", + "answer_gt": "[0.0, 1.06]", + "answer_gt_word": "dog bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11595 + }, + { + "path": "Y5-xKnAigQnA.wav", + "question": "At what time points is 'a dog is barking' fully present in the recording?", + "choice_a": "[0.36, 10.0]", + "choice_b": "[1.16, 3.51]", + "choice_c": "[1.83, 8.68]", + "choice_d": "[2.18, 9.02]", + "answer_gt": "[0.36, 10.0]", + "answer_gt_word": "a dog is barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11596 + }, + { + "path": "Y55NFpupN4FM.wav", + "question": "Where is 'a dog runs' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 0.33]", + "choice_b": "[0.82, 2.02]", + "choice_c": "[2.16, 7.9]", + "choice_d": "[8.78, 9.87]", + "answer_gt": "[2.16, 7.9]", + "answer_gt_word": "a dog runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11597 + }, + { + "path": "Y56wmOXDjjZY.wav", + "question": "Which sections of the sound contain 'a man speaks' in their entirety?", + "choice_a": "[0.0, 1.28]", + "choice_b": "[3.87, 6.13]", + "choice_c": "[6.74, 10.0]", + "choice_d": "[1.54, 3.22]", + "answer_gt": "[1.54, 3.22]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11598 + }, + { + "path": "Y57NefYJLTE4.wav", + "question": "Where do we hear 'a police siren amid general vehicle noises' in full within this sound recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.67, 6.84]", + "choice_c": "[1.6, 7.22]", + "choice_d": "[4.1, 8.59]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a police siren amid general vehicle noises", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11599 + }, + { + "path": "Y5ALyWjiYMEo.wav", + "question": "What ranges in seconds capture 'dog panting' in full?", + "choice_a": "[0.25, 6.75]", + "choice_b": "[2.56, 3.46]", + "choice_c": "[4.42, 4.87]", + "choice_d": "[9.04, 10.0]", + "answer_gt": "[0.25, 6.75]", + "answer_gt_word": "dog panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11600 + }, + { + "path": "Y5CpDdfP4Yes.wav", + "question": "At what seconds do the full iterations of 'dog panting' start and end?", + "choice_a": "[0.25, 6.75]", + "choice_b": "[1.58, 4.27]", + "choice_c": "[5.05, 9.92]", + "choice_d": "[4.47, 4.69]", + "answer_gt": "[0.25, 6.75]", + "answer_gt_word": "dog panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11601 + }, + { + "path": "Y5EFYHBn7660.wav", + "question": "Which sections of the sound contain 'constant water flowing in background' in their entirety?", + "choice_a": "[0.0, 1.36]", + "choice_b": "[5.58, 6.11]", + "choice_c": "[8.6, 9.25]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "constant water flowing in background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11602 + }, + { + "path": "Y5FoC4i8ZbfQ.wav", + "question": "What are the precise second intervals of 'nails scratch on tile' in this clip?", + "choice_a": "[0.0, 3.8]", + "choice_b": "[9.7, 10.0]", + "choice_c": "[2.74, 4.01]", + "choice_d": "[4.21, 6.21]", + "answer_gt": "[2.74, 4.01]", + "answer_gt_word": "nails scratch on tile", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11603 + }, + { + "path": "Y5GlL2iqv-I4.wav", + "question": "How long does 'water waves crash' last each time it appears in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.61, 3.16]", + "choice_c": "[6.45, 8.53]", + "choice_d": "[0.26, 1.44]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "water waves crash", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11604 + }, + { + "path": "Y5IdhdVtHUVY.wav", + "question": "Where is 'a helicopter over the waves' found in its entirety within the audio's timeline?", + "choice_a": "[0.12, 8.93]", + "choice_b": "[3.32, 4.77]", + "choice_c": "[6.13, 6.63]", + "choice_d": "[7.18, 8.09]", + "answer_gt": "[0.12, 8.93]", + "answer_gt_word": "a helicopter over the waves", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11605 + }, + { + "path": "Y5QYvAqQaDoE.wav", + "question": "What are the full time segments when 'a propeller plane is accelerating in the back' appears in this sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.12, 2.04]", + "choice_c": "[0.69, 7.08]", + "choice_d": "[3.48, 9.6]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a propeller plane is accelerating in the back", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11606 + }, + { + "path": "Y5RklOICz5eo.wav", + "question": "What timestamps mark the full occurrences of 'music plays'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.48, 10.0]", + "choice_c": "[5.46, 7.19]", + "choice_d": "[6.43, 9.28]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "music plays", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11607 + }, + { + "path": "Y5UYqZWCvud8.wav", + "question": "List the periods when 'hissing is present' is completely played in the sound.", + "choice_a": "[0.16, 1.62]", + "choice_b": "[3.74, 5.63]", + "choice_c": "[0.0, 9.65]", + "choice_d": "[1.0, 2.11]", + "answer_gt": "[0.0, 9.65]", + "answer_gt_word": "hissing is present", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11608 + }, + { + "path": "Y5YXlfyojVcg.wav", + "question": "What are the intervals in seconds of each 'man talking' appearance?", + "choice_a": "[0.0, 1.73]", + "choice_b": "[2.92, 10.0]", + "choice_c": "[2.96, 10.02]", + "choice_d": "[1.28, 6.08]", + "answer_gt": "[0.0, 1.73]", + "answer_gt_word": "man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11609 + }, + { + "path": "Y5fihLVMJdtc.wav", + "question": "What are the specific start and stop times of 'man talking' in the sound?", + "choice_a": "[0.23, 0.65]", + "choice_b": "[3.29, 9.15]", + "choice_c": "[0.37, 1.18]", + "choice_d": "[0.0, 1.73]", + "answer_gt": "[0.0, 1.73]", + "answer_gt_word": "man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11610 + }, + { + "path": "Y5hxPAI1jG3s.wav", + "question": "Provide the start and end times of 'an animal barks' within this audio.", + "choice_a": "[1.85, 2.42]", + "choice_b": "[1.81, 4.35]", + "choice_c": "[0.0, 0.43]", + "choice_d": "[1.32, 1.87]", + "answer_gt": "[1.85, 2.42]", + "answer_gt_word": "an animal barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11611 + }, + { + "path": "Y5jQKX0Kmv6c.wav", + "question": "What are the full time segments when 'a race car is accelerating away' appears in this sound?", + "choice_a": "[0.02, 10.0]", + "choice_b": "[3.3, 3.79]", + "choice_c": "[4.46, 4.96]", + "choice_d": "[7.27, 7.99]", + "answer_gt": "[0.02, 10.0]", + "answer_gt_word": "a race car is accelerating away", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11612 + }, + { + "path": "Y5mIHonRMBE4.wav", + "question": "At what time points is 'man speaking' fully present in the recording?", + "choice_a": "[0.0, 1.74]", + "choice_b": "[1.72, 3.18]", + "choice_c": "[2.08, 4.89]", + "choice_d": "[2.87, 5.98]", + "answer_gt": "[0.0, 1.74]", + "answer_gt_word": "man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11613 + }, + { + "path": "Y5rdR6yhKU1A.wav", + "question": "Capture the timeline where 'car accelerating at high speed' is fully included in the sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 0.87]", + "choice_c": "[2.88, 7.89]", + "choice_d": "[5.97, 9.55]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "car accelerating at high speed", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11614 + }, + { + "path": "Y5sAeoy_rmAg.wav", + "question": "Enumerate the second-based occurrences of 'strong wind is blowing' in full.", + "choice_a": "[0.0, 10.01]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[1.82, 4.87]", + "choice_d": "[4.34, 5.39]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "strong wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11615 + }, + { + "path": "Y5uWwl9DrpW0.wav", + "question": "How many seconds does 'strong wind is blowing' span in each occurrence?", + "choice_a": "[0.19, 0.49]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[6.48, 6.75]", + "choice_d": "[7.47, 7.81]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "strong wind is blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11616 + }, + { + "path": "Y64FeLMlvRK8.wav", + "question": "Locate every segment where 'rustling' is completely audible.", + "choice_a": "[0.0, 1.84]", + "choice_b": "[2.53, 3.12]", + "choice_c": "[5.44, 5.9]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11617 + }, + { + "path": "Y656Qt18XG68.wav", + "question": "How long does 'a few dogs are barking and whimper' last each time it appears in the audio?", + "choice_a": "[3.79, 9.85]", + "choice_b": "[2.33, 3.5]", + "choice_c": "[2.63, 3.95]", + "choice_d": "[2.79, 9.27]", + "answer_gt": "[3.79, 9.85]", + "answer_gt_word": "a few dogs are barking and whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11618 + }, + { + "path": "Y66YGQBFkhEY.wav", + "question": "In what time spans is 'a few dogs are barking and whimper' completely audible in this recording?", + "choice_a": "[3.79, 9.85]", + "choice_b": "[1.03, 1.4]", + "choice_c": "[1.82, 2.21]", + "choice_d": "[3.14, 7.03]", + "answer_gt": "[3.79, 9.85]", + "answer_gt_word": "a few dogs are barking and whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11619 + }, + { + "path": "Y66cUgRJThiw.wav", + "question": "How long does 'a few dogs are barking and whimper' last each time it appears in the audio?", + "choice_a": "[2.82, 3.51]", + "choice_b": "[7.71, 8.98]", + "choice_c": "[0.5, 0.75]", + "choice_d": "[3.79, 9.85]", + "answer_gt": "[3.79, 9.85]", + "answer_gt_word": "a few dogs are barking and whimper", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11620 + }, + { + "path": "Y66h0PEq4EwU.wav", + "question": "List the specific seconds of 'a puppy cries' appearances.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.17, 3.06]", + "choice_c": "[1.22, 4.08]", + "choice_d": "[2.92, 8.03]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a puppy cries", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11621 + }, + { + "path": "Y67NYH3dKEd4.wav", + "question": "At what time intervals does 'an animal makes a whimpering noise' occur in the sound file?", + "choice_a": "[0.0, 3.87]", + "choice_b": "[0.17, 6.38]", + "choice_c": "[5.18, 7.56]", + "choice_d": "[6.18, 9.33]", + "answer_gt": "[0.0, 3.87]", + "answer_gt_word": "an animal makes a whimpering noise", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11622 + }, + { + "path": "Y6IzqjDGIDao.wav", + "question": "What second marks correspond to the entire span of 'something rustles'?", + "choice_a": "[1.85, 2.81]", + "choice_b": "[4.72, 5.21]", + "choice_c": "[8.36, 9.14]", + "choice_d": "[5.18, 10.0]", + "answer_gt": "[5.18, 10.0]", + "answer_gt_word": "something rustles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11623 + }, + { + "path": "Y6Jwm7Uz2K9k.wav", + "question": "What are the boundaries in seconds for each 'something rustles' presence?", + "choice_a": "[5.18, 10.0]", + "choice_b": "[3.99, 5.68]", + "choice_c": "[7.16, 9.3]", + "choice_d": "[0.64, 3.03]", + "answer_gt": "[5.18, 10.0]", + "answer_gt_word": "something rustles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11624 + }, + { + "path": "Y6KRIQDFUeDs.wav", + "question": "What are the intervals in seconds of each 'something rustles' appearance?", + "choice_a": "[5.18, 10.0]", + "choice_b": "[0.98, 1.45]", + "choice_c": "[2.0, 2.45]", + "choice_d": "[3.64, 4.08]", + "answer_gt": "[5.18, 10.0]", + "answer_gt_word": "something rustles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11625 + }, + { + "path": "Y6M6bhcnYLLk.wav", + "question": "List the specific seconds of 'a dog panting' appearances.", + "choice_a": "[0.23, 9.76]", + "choice_b": "[3.14, 3.94]", + "choice_c": "[0.45, 7.33]", + "choice_d": "[7.25, 7.82]", + "answer_gt": "[0.23, 9.76]", + "answer_gt_word": "a dog panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11626 + }, + { + "path": "Y6Q7JIacR29o.wav", + "question": "Document the full appearances of 'a train horn blows' in terms of seconds.", + "choice_a": "[1.07, 10.0]", + "choice_b": "[0.29, 1.07]", + "choice_c": "[0.97, 7.03]", + "choice_d": "[2.21, 9.73]", + "answer_gt": "[1.07, 10.0]", + "answer_gt_word": "a train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11627 + }, + { + "path": "Y6T2jEKesKXY.wav", + "question": "How are the complete segments of 'a train horn blows' timed in this audio?", + "choice_a": "[0.0, 1.54]", + "choice_b": "[2.68, 6.74]", + "choice_c": "[1.07, 10.0]", + "choice_d": "[1.13, 3.9]", + "answer_gt": "[1.07, 10.0]", + "answer_gt_word": "a train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11628 + }, + { + "path": "Y6YKDBuntOF0.wav", + "question": "Determine the full durations of 'a train horn blows' in the audio file.", + "choice_a": "[1.5, 1.89]", + "choice_b": "[2.89, 3.18]", + "choice_c": "[1.07, 10.0]", + "choice_d": "[9.14, 9.8]", + "answer_gt": "[1.07, 10.0]", + "answer_gt_word": "a train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11629 + }, + { + "path": "Y6YWK8S1thvk.wav", + "question": "What time segments delineate the presence of 'the wind roars'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.82, 4.25]", + "choice_d": "[1.95, 5.78]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "the wind roars", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11630 + }, + { + "path": "Y6cS0FsUM-cQ.wav", + "question": "How long does 'an adult male speaks' last each time it appears in the audio?", + "choice_a": "[0.01, 0.35]", + "choice_b": "[8.01, 9.41]", + "choice_c": "[0.21, 2.17]", + "choice_d": "[3.46, 10.02]", + "answer_gt": "[0.01, 0.35]", + "answer_gt_word": "an adult male speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11631 + }, + { + "path": "Y6cWIWgtMLjg.wav", + "question": "Identify the periods when 'a person laughs' fully resonates in the sound.", + "choice_a": "[1.4, 2.84]", + "choice_b": "[3.68, 7.37]", + "choice_c": "[7.87, 10.0]", + "choice_d": "[2.97, 3.56]", + "answer_gt": "[1.4, 2.84]", + "answer_gt_word": "a person laughs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11632 + }, + { + "path": "Y6eHnIvL0N7M.wav", + "question": "What are the precise second intervals of 'male speech' in this clip?", + "choice_a": "[0.32, 4.07]", + "choice_b": "[9.01, 9.3]", + "choice_c": "[4.75, 8.16]", + "choice_d": "[0.0, 3.5]", + "answer_gt": "[4.75, 8.16]", + "answer_gt_word": "male speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11633 + }, + { + "path": "Y6h4hZiywOxE.wav", + "question": "At which seconds does 'vehicle engine being shut off' fully register in the sound?", + "choice_a": "[0.0, 2.0]", + "choice_b": "[0.0, 1.26]", + "choice_c": "[1.99, 4.89]", + "choice_d": "[5.53, 10.0]", + "answer_gt": "[0.0, 2.0]", + "answer_gt_word": "vehicle engine being shut off", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11634 + }, + { + "path": "Y6jFl0KW-S84.wav", + "question": "What time segments delineate the presence of 'a door is opened'?", + "choice_a": "[0.19, 1.01]", + "choice_b": "[4.2, 5.93]", + "choice_c": "[6.9, 10.0]", + "choice_d": "[4.77, 5.6]", + "answer_gt": "[0.19, 1.01]", + "answer_gt_word": "a door is opened", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11635 + }, + { + "path": "Y6nPirCzxDV8.wav", + "question": "Capture the timeline where 'a door is opened' is fully included in the sound.", + "choice_a": "[0.38, 2.68]", + "choice_b": "[3.72, 5.11]", + "choice_c": "[0.19, 1.01]", + "choice_d": "[0.39, 2.68]", + "answer_gt": "[0.19, 1.01]", + "answer_gt_word": "a door is opened", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11636 + }, + { + "path": "Y6nr3SWNml9A.wav", + "question": "Capture the timeline where 'a large engine is idling' is fully included in the sound.", + "choice_a": "[0.0, 9.09]", + "choice_b": "[0.0, 1.59]", + "choice_c": "[7.58, 9.05]", + "choice_d": "[6.64, 9.05]", + "answer_gt": "[0.0, 9.09]", + "answer_gt_word": "a large engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11637 + }, + { + "path": "Y6onVPeqqNsk.wav", + "question": "What are the precise second intervals of 'a large engine is idling' in this clip?", + "choice_a": "[0.0, 2.69]", + "choice_b": "[3.39, 4.23]", + "choice_c": "[5.58, 9.55]", + "choice_d": "[0.0, 9.09]", + "answer_gt": "[0.0, 9.09]", + "answer_gt_word": "a large engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11638 + }, + { + "path": "Y6qxcYCOkw8I.wav", + "question": "At what seconds do the full iterations of 'a cat meow' start and end?", + "choice_a": "[0.0, 0.23]", + "choice_b": "[0.93, 1.65]", + "choice_c": "[2.57, 3.45]", + "choice_d": "[5.98, 6.96]", + "answer_gt": "[5.98, 6.96]", + "answer_gt_word": "a cat meow", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11639 + }, + { + "path": "Y6sFsyoK8Nn0.wav", + "question": "In what time spans is 'a cat meow' completely audible in this recording?", + "choice_a": "[0.0, 1.3]", + "choice_b": "[1.93, 4.32]", + "choice_c": "[4.62, 5.43]", + "choice_d": "[5.98, 6.96]", + "answer_gt": "[5.98, 6.96]", + "answer_gt_word": "a cat meow", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11640 + }, + { + "path": "Y71iXgJIvlTg.wav", + "question": "What are the full time segments when 'a woman is speaking gently' appears in this sound?", + "choice_a": "[0.26, 0.74]", + "choice_b": "[2.85, 3.66]", + "choice_c": "[3.96, 5.66]", + "choice_d": "[4.41, 6.77]", + "answer_gt": "[4.41, 6.77]", + "answer_gt_word": "a woman is speaking gently", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11641 + }, + { + "path": "Y720-k49ehJA.wav", + "question": "Enumerate the second-based occurrences of 'a whimpering dog' in full.", + "choice_a": "[3.2, 4.71]", + "choice_b": "[8.8, 9.01]", + "choice_c": "[7.52, 7.87]", + "choice_d": "[3.7, 5.76]", + "answer_gt": "[7.52, 7.87]", + "answer_gt_word": "a whimpering dog", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11642 + }, + { + "path": "Y72J8fjETl6U.wav", + "question": "In what segments can 'laughing' be heard from start to finish?", + "choice_a": "[0.74, 3.96]", + "choice_b": "[6.9, 8.1]", + "choice_c": "[8.92, 9.16]", + "choice_d": "[4.13, 6.12]", + "answer_gt": "[4.13, 6.12]", + "answer_gt_word": "laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11643 + }, + { + "path": "Y736sQyNeUpc.wav", + "question": "At what seconds do the full iterations of 'wind blowing' start and end?", + "choice_a": "[0.01, 10.0]", + "choice_b": "[9.45, 10.0]", + "choice_c": "[5.04, 5.4]", + "choice_d": "[8.83, 8.96]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "wind blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11644 + }, + { + "path": "Y73phKdahss8.wav", + "question": "How many seconds does each full presence of 'wind blowing' cover?", + "choice_a": "[0.0, 1.38]", + "choice_b": "[3.07, 5.84]", + "choice_c": "[0.01, 10.0]", + "choice_d": "[0.97, 5.82]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "wind blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11645 + }, + { + "path": "Y78nshhXgM6E.wav", + "question": "What are all the complete time segments of 'man talking' in this sound?", + "choice_a": "[0.0, 3.1]", + "choice_b": "[3.89, 5.83]", + "choice_c": "[4.89, 6.03]", + "choice_d": "[9.2, 10.0]", + "answer_gt": "[4.89, 6.03]", + "answer_gt_word": "man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11646 + }, + { + "path": "Y7BUwVfKkw_w.wav", + "question": "Identify the periods when 'man talking' fully resonates in the sound.", + "choice_a": "[4.89, 6.03]", + "choice_b": "[4.52, 5.45]", + "choice_c": "[6.81, 7.83]", + "choice_d": "[2.72, 4.38]", + "answer_gt": "[4.89, 6.03]", + "answer_gt_word": "man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11647 + }, + { + "path": "Y7IzG88ix5Wc.wav", + "question": "Document the full appearances of 'slows down' in terms of seconds.", + "choice_a": "[2.6, 9.08]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.42, 3.58]", + "choice_d": "[2.47, 9.1]", + "answer_gt": "[2.6, 9.08]", + "answer_gt_word": "slows down", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11648 + }, + { + "path": "Y7OAWoKiUEXc.wav", + "question": "What are the intervals in seconds of each 'an emergency vehicle blares its horn' appearance?", + "choice_a": "[0.0, 10.01]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[0.15, 8.26]", + "choice_d": "[7.72, 9.7]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "an emergency vehicle blares its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11649 + }, + { + "path": "Y7S3Yf3QBvzk.wav", + "question": "What are the intervals in seconds of each 'people talk' appearance?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.42, 0.77]", + "choice_c": "[1.66, 3.05]", + "choice_d": "[8.12, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "people talk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11650 + }, + { + "path": "Y7WP4vfMgZsc.wav", + "question": "What are the distinct time ranges of 'distant murmuring' within the audio?", + "choice_a": "[1.94, 5.57]", + "choice_b": "[1.54, 2.31]", + "choice_c": "[4.69, 4.98]", + "choice_d": "[5.85, 7.0]", + "answer_gt": "[1.94, 5.57]", + "answer_gt_word": "distant murmuring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11651 + }, + { + "path": "Y7ZD1_tPZ1qE.wav", + "question": "At what points does 'an engine idles' exist completely in the timeline?", + "choice_a": "[4.6, 10.0]", + "choice_b": "[0.02, 4.48]", + "choice_c": "[0.58, 8.14]", + "choice_d": "[1.06, 9.89]", + "answer_gt": "[4.6, 10.0]", + "answer_gt_word": "an engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11652 + }, + { + "path": "Y7aI9LqS4jK4.wav", + "question": "At what timestamps does the complete 'an engine idles' occur?", + "choice_a": "[4.6, 10.0]", + "choice_b": "[1.54, 2.14]", + "choice_c": "[4.63, 10.0]", + "choice_d": "[0.0, 1.76]", + "answer_gt": "[4.6, 10.0]", + "answer_gt_word": "an engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11653 + }, + { + "path": "Y7b3s4rfWF50.wav", + "question": "What are the boundaries in seconds for each 'dogs bark' presence?", + "choice_a": "[0.0, 1.14]", + "choice_b": "[2.44, 3.17]", + "choice_c": "[3.98, 5.66]", + "choice_d": "[8.6, 8.8]", + "answer_gt": "[8.6, 8.8]", + "answer_gt_word": "dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11654 + }, + { + "path": "Y7i8tr14DaOI.wav", + "question": "Capture the timeline where 'dogs bark' is fully included in the sound.", + "choice_a": "[0.67, 1.0]", + "choice_b": "[8.6, 8.8]", + "choice_c": "[6.35, 7.1]", + "choice_d": "[9.48, 9.82]", + "answer_gt": "[8.6, 8.8]", + "answer_gt_word": "dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11655 + }, + { + "path": "Y7li4oPDlkvo.wav", + "question": "What are the full time segments when 'a bell is dinging' appears in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[4.58, 10.0]", + "choice_c": "[0.52, 4.93]", + "choice_d": "[0.87, 8.18]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a bell is dinging", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11656 + }, + { + "path": "Y7mQepeQImgY.wav", + "question": "Which sections of the sound contain 'railroad train hitting the track' in their entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.01, 3.64]", + "choice_c": "[4.88, 6.36]", + "choice_d": "[7.78, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "railroad train hitting the track", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11657 + }, + { + "path": "Y7sHIi7WZkDA.wav", + "question": "What time segments delineate the presence of 'rustling occurs'?", + "choice_a": "[0.0, 0.16]", + "choice_b": "[8.35, 8.61]", + "choice_c": "[0.0, 9.34]", + "choice_d": "[1.46, 2.26]", + "answer_gt": "[0.0, 9.34]", + "answer_gt_word": "rustling occurs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11658 + }, + { + "path": "Y7wyV4mU-E4k.wav", + "question": "How are the complete segments of 'a television is playing something in the background' timed in this audio?", + "choice_a": "[0.01, 10.0]", + "choice_b": "[1.93, 2.71]", + "choice_c": "[3.45, 4.1]", + "choice_d": "[5.36, 6.28]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "a television is playing something in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11659 + }, + { + "path": "Y7xjWD6j5-RY.wav", + "question": "Where does 'an aircraft engine is humming' appear from beginning to end in the sound?", + "choice_a": "[2.82, 5.89]", + "choice_b": "[8.11, 10.0]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[1.11, 8.95]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an aircraft engine is humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11660 + }, + { + "path": "Y84NFbeQy5CU.wav", + "question": "Identify the complete instances of 'a motorcycle is racing around' within the sound timeline.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.7, 5.24]", + "choice_c": "[8.36, 10.0]", + "choice_d": "[0.7, 9.34]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a motorcycle is racing around", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11661 + }, + { + "path": "Y88id3bEvC4s.wav", + "question": "At which seconds does 'a kid talks' fully register in the sound?", + "choice_a": "[0.95, 1.95]", + "choice_b": "[3.99, 4.52]", + "choice_c": "[4.62, 5.99]", + "choice_d": "[4.8, 5.54]", + "answer_gt": "[4.62, 5.99]", + "answer_gt_word": "a kid talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11662 + }, + { + "path": "Y8InQJ0Nng0M.wav", + "question": "What time segments delineate the presence of 'a kid talks'?", + "choice_a": "[3.12, 3.79]", + "choice_b": "[4.5, 5.31]", + "choice_c": "[0.03, 2.19]", + "choice_d": "[4.62, 5.99]", + "answer_gt": "[4.62, 5.99]", + "answer_gt_word": "a kid talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11663 + }, + { + "path": "Y8Rb3I8SyT1U.wav", + "question": "Which sections of the sound contain 'an emergency motor vehicle siren blows' in their entirety?", + "choice_a": "[0.01, 3.1]", + "choice_b": "[3.01, 10.0]", + "choice_c": "[3.01, 10.0]", + "choice_d": "[3.08, 10.0]", + "answer_gt": "[0.01, 3.1]", + "answer_gt_word": "an emergency motor vehicle siren blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11664 + }, + { + "path": "Y8WbUPEYjo1o.wav", + "question": "What are the boundaries in seconds for each 'loud engine of vehicle repeatedly driving by' presence?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.98, 8.28]", + "choice_c": "[3.59, 8.62]", + "choice_d": "[8.05, 9.7]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud engine of vehicle repeatedly driving by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11665 + }, + { + "path": "Y8ZmSpxseOtI.wav", + "question": "How long does 'loud engine of vehicle repeatedly driving by' last each time it appears in the audio?", + "choice_a": "[4.17, 4.56]", + "choice_b": "[5.31, 5.65]", + "choice_c": "[7.73, 8.3]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud engine of vehicle repeatedly driving by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11666 + }, + { + "path": "Y8fVq7MMxhVc.wav", + "question": "At what timestamps does the complete 'loud engine of vehicle repeatedly driving by' occur?", + "choice_a": "[0.81, 1.31]", + "choice_b": "[2.79, 3.75]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[7.06, 9.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "loud engine of vehicle repeatedly driving by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11667 + }, + { + "path": "Y8g8UfyxJ3qE.wav", + "question": "What are all the complete time segments of 'an engine takes a few seconds to start' in this sound?", + "choice_a": "[0.01, 3.01]", + "choice_b": "[1.43, 9.99]", + "choice_c": "[4.14, 6.55]", + "choice_d": "[5.78, 8.68]", + "answer_gt": "[0.01, 3.01]", + "answer_gt_word": "an engine takes a few seconds to start", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11668 + }, + { + "path": "Y8kDguJE2oC8.wav", + "question": "Identify the periods when 'a device is vibrating' fully resonates in the sound.", + "choice_a": "[0.3, 1.47]", + "choice_b": "[1.9, 5.29]", + "choice_c": "[6.15, 7.0]", + "choice_d": "[0.0, 9.22]", + "answer_gt": "[0.0, 9.22]", + "answer_gt_word": "a device is vibrating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11669 + }, + { + "path": "Y8lhdsS_xrvg.wav", + "question": "At what time intervals does 'a motor is running' occur in the sound file?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.47, 3.55]", + "choice_c": "[1.65, 3.58]", + "choice_d": "[2.68, 7.28]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a motor is running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11670 + }, + { + "path": "Y8naj2THt5C0.wav", + "question": "At what timestamps does the complete 'man' occur?", + "choice_a": "[0.55, 2.04]", + "choice_b": "[3.83, 5.04]", + "choice_c": "[5.53, 7.99]", + "choice_d": "[7.52, 8.82]", + "answer_gt": "[5.53, 7.99]", + "answer_gt_word": "man", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11671 + }, + { + "path": "Y8tskw5JW2d0.wav", + "question": "Identify the complete instances of 'man' within the sound timeline.", + "choice_a": "[1.0, 2.15]", + "choice_b": "[4.66, 5.08]", + "choice_c": "[5.53, 7.99]", + "choice_d": "[2.19, 4.68]", + "answer_gt": "[5.53, 7.99]", + "answer_gt_word": "man", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11672 + }, + { + "path": "Y8ycflE3dIHw.wav", + "question": "At what points does 'short honk' exist completely in the timeline?", + "choice_a": "[5.86, 6.68]", + "choice_b": "[4.02, 7.25]", + "choice_c": "[4.86, 9.02]", + "choice_d": "[4.86, 9.66]", + "answer_gt": "[5.86, 6.68]", + "answer_gt_word": "short honk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11673 + }, + { + "path": "Y9Dz4Sisz-Z4.wav", + "question": "What time segments delineate the presence of 'a large engine idling'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[6.13, 6.43]", + "choice_c": "[6.68, 7.96]", + "choice_d": "[7.05, 7.96]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a large engine idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11674 + }, + { + "path": "Y9FnSoyIhUGI.wav", + "question": "Which sections of the sound contain 'an engine idling' in their entirety?", + "choice_a": "[0.0, 8.87]", + "choice_b": "[6.26, 6.78]", + "choice_c": "[9.65, 10.0]", + "choice_d": "[1.14, 9.56]", + "answer_gt": "[0.0, 8.87]", + "answer_gt_word": "an engine idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11675 + }, + { + "path": "Y9KuT6SOwobY.wav", + "question": "Provide the start and end times of 'a man speaking' within this audio.", + "choice_a": "[0.06, 0.85]", + "choice_b": "[0.94, 1.85]", + "choice_c": "[0.31, 4.66]", + "choice_d": "[4.47, 9.03]", + "answer_gt": "[0.06, 0.85]", + "answer_gt_word": "a man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11676 + }, + { + "path": "Y9LQEZJPNVpw.wav", + "question": "What timestamps mark the full occurrences of 'a horn blowing'?", + "choice_a": "[0.0, 9.79]", + "choice_b": "[0.43, 4.1]", + "choice_c": "[0.98, 5.53]", + "choice_d": "[2.98, 6.9]", + "answer_gt": "[0.0, 9.79]", + "answer_gt_word": "a horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11677 + }, + { + "path": "Y9ZDBsJjzNOY.wav", + "question": "How is 'pigeons coo loudly and incessantly' distributed over the sound's duration?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[5.65, 7.08]", + "choice_d": "[7.07, 8.93]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "pigeons coo loudly and incessantly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11678 + }, + { + "path": "Y9jDfLBPWwZM.wav", + "question": "What is the range in seconds for each full 'pigeons coo loudly and incessantly' in the recording?", + "choice_a": "[0.0, 2.36]", + "choice_b": "[7.23, 10.0]", + "choice_c": "[2.35, 4.17]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "pigeons coo loudly and incessantly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11679 + }, + { + "path": "Y9o3monl8eBU.wav", + "question": "How many seconds does each full presence of 'pigeons coo loudly and incessantly' cover?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[2.22, 8.55]", + "choice_c": "[0.0, 1.69]", + "choice_d": "[2.24, 8.58]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "pigeons coo loudly and incessantly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11680 + }, + { + "path": "Y9qj7l2LE2PM.wav", + "question": "What time segments delineate the presence of 'light wind'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[7.16, 8.05]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[7.76, 9.3]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "light wind", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11681 + }, + { + "path": "Y9qwuWhM2C8Q.wav", + "question": "Where do we hear 'light wind' in full within this sound recording?", + "choice_a": "[0.0, 1.13]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[3.45, 9.98]", + "choice_d": "[1.34, 1.91]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "light wind", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11682 + }, + { + "path": "YA-7kQF3QN54.wav", + "question": "Identify the periods when 'rustling' fully resonates in the sound.", + "choice_a": "[0.0, 7.72]", + "choice_b": "[0.62, 2.61]", + "choice_c": "[5.01, 7.23]", + "choice_d": "[4.89, 8.12]", + "answer_gt": "[0.0, 7.72]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11683 + }, + { + "path": "YA02FI-n7uDk.wav", + "question": "What time segments showcase 'engine being started' in its entirety?", + "choice_a": "[0.0, 2.37]", + "choice_b": "[2.44, 10.02]", + "choice_c": "[0.48, 3.65]", + "choice_d": "[3.39, 9.62]", + "answer_gt": "[0.0, 2.37]", + "answer_gt_word": "engine being started", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11684 + }, + { + "path": "YA21eA2M6cuM.wav", + "question": "Identify the complete instances of 'radio static' within the sound timeline.", + "choice_a": "[0.15, 10.0]", + "choice_b": "[1.91, 4.51]", + "choice_c": "[5.14, 8.93]", + "choice_d": "[9.6, 10.0]", + "answer_gt": "[0.15, 10.0]", + "answer_gt_word": "radio static", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11685 + }, + { + "path": "YAfrV3HG7JFw.wav", + "question": "Determine the full durations of 'a man speaks' in the audio file.", + "choice_a": "[0.0, 4.07]", + "choice_b": "[0.73, 4.72]", + "choice_c": "[0.99, 7.36]", + "choice_d": "[2.33, 8.17]", + "answer_gt": "[0.0, 4.07]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11686 + }, + { + "path": "YAlzK5RtVIyc.wav", + "question": "Where is 'heavy , slow rainfall' consistently heard throughout the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.79, 7.72]", + "choice_c": "[4.46, 8.58]", + "choice_d": "[7.21, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "heavy , slow rainfall", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11687 + }, + { + "path": "YAp8Eb8blOQA.wav", + "question": "What are the distinct time ranges of 'an engine idling' within the audio?", + "choice_a": "[0.0, 9.99]", + "choice_b": "[0.52, 1.04]", + "choice_c": "[6.35, 7.42]", + "choice_d": "[4.49, 6.12]", + "answer_gt": "[0.0, 9.99]", + "answer_gt_word": "an engine idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11688 + }, + { + "path": "YAtmv7HYs-uk.wav", + "question": "Enumerate the second-based occurrences of 'an engine running quietly' in full.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.93, 3.49]", + "choice_c": "[4.58, 5.93]", + "choice_d": "[8.95, 9.79]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine running quietly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11689 + }, + { + "path": "YBDXGM5qwT_w.wav", + "question": "Document the full appearances of 'clattering occurs' in terms of seconds.", + "choice_a": "[0.0, 0.32]", + "choice_b": "[1.97, 2.3]", + "choice_c": "[8.02, 9.06]", + "choice_d": "[8.72, 9.04]", + "answer_gt": "[0.0, 0.32]", + "answer_gt_word": "clattering occurs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11690 + }, + { + "path": "YBLTscl4P0RI.wav", + "question": "Where is 'a loud horn honking' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 4.25]", + "choice_b": "[4.41, 10.01]", + "choice_c": "[0.43, 3.64]", + "choice_d": "[0.78, 9.62]", + "answer_gt": "[0.0, 4.25]", + "answer_gt_word": "a loud horn honking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11691 + }, + { + "path": "YBML3g2YRXt0.wav", + "question": "What are the precise second intervals of 'a man speaks' in this clip?", + "choice_a": "[1.15, 1.44]", + "choice_b": "[2.06, 4.1]", + "choice_c": "[0.18, 4.76]", + "choice_d": "[3.17, 6.63]", + "answer_gt": "[1.15, 1.44]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11692 + }, + { + "path": "YBXxlqaDvdaA.wav", + "question": "In what segments can 'wind blowing hard over microphone' be heard from start to finish?", + "choice_a": "[0.82, 2.58]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[6.39, 6.79]", + "choice_d": "[8.27, 9.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind blowing hard over microphone", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11693 + }, + { + "path": "YC7yZovKH0Jg.wav", + "question": "In what time spans is 'wind blows' completely audible in this recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.82, 6.03]", + "choice_d": "[4.89, 9.17]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11694 + }, + { + "path": "YCh0LMmhBUg4.wav", + "question": "Identify the periods when 'a child speaks' fully resonates in the sound.", + "choice_a": "[0.0, 0.77]", + "choice_b": "[1.93, 4.19]", + "choice_c": "[5.69, 7.88]", + "choice_d": "[2.96, 4.04]", + "answer_gt": "[2.96, 4.04]", + "answer_gt_word": "a child speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11695 + }, + { + "path": "YDMtUvX8y87M.wav", + "question": "What time segments showcase 'a child speaks' in its entirety?", + "choice_a": "[2.96, 4.04]", + "choice_b": "[2.83, 3.46]", + "choice_c": "[5.65, 5.96]", + "choice_d": "[6.98, 7.29]", + "answer_gt": "[2.96, 4.04]", + "answer_gt_word": "a child speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11696 + }, + { + "path": "YDcM_wehXfeg.wav", + "question": "At which time markers does 'a child speaks' fully unfold in the sound?", + "choice_a": "[2.96, 4.04]", + "choice_b": "[4.37, 5.86]", + "choice_c": "[0.01, 2.01]", + "choice_d": "[2.65, 4.4]", + "answer_gt": "[2.96, 4.04]", + "answer_gt_word": "a child speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11697 + }, + { + "path": "YE3sNwZlVRv4.wav", + "question": "Identify the complete instances of 'railroad crossing signals ringing' within the sound timeline.", + "choice_a": "[0.01, 8.62]", + "choice_b": "[0.18, 3.26]", + "choice_c": "[4.45, 8.06]", + "choice_d": "[5.29, 9.34]", + "answer_gt": "[0.01, 8.62]", + "answer_gt_word": "railroad crossing signals ringing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11698 + }, + { + "path": "YECPfahcqU4k.wav", + "question": "What are all the complete time segments of 'railroad warning signals beeping' in this sound?", + "choice_a": "[0.01, 10.01]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.74, 2.46]", + "choice_d": "[5.66, 8.42]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "railroad warning signals beeping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11699 + }, + { + "path": "YEONfjVbpIuM.wav", + "question": "Identify the complete instances of 'wind noise sounds' within the sound timeline.", + "choice_a": "[0.0, 3.47]", + "choice_b": "[3.58, 10.01]", + "choice_c": "[3.79, 6.64]", + "choice_d": "[6.58, 9.01]", + "answer_gt": "[0.0, 3.47]", + "answer_gt_word": "wind noise sounds", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11700 + }, + { + "path": "YETXMll5dC8Q.wav", + "question": "Define the time intervals that fully contain 'an idle vehicle engine running' in this sound.", + "choice_a": "[0.0, 9.25]", + "choice_b": "[2.33, 4.69]", + "choice_c": "[2.61, 6.5]", + "choice_d": "[3.66, 9.05]", + "answer_gt": "[0.0, 9.25]", + "answer_gt_word": "an idle vehicle engine running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11701 + }, + { + "path": "YFcD3ukcMbCc.wav", + "question": "What are the intervals in seconds of each 'splashing water' appearance?", + "choice_a": "[0.0, 0.78]", + "choice_b": "[2.54, 3.15]", + "choice_c": "[3.83, 10.0]", + "choice_d": "[3.17, 10.0]", + "answer_gt": "[3.17, 10.0]", + "answer_gt_word": "splashing water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11702 + }, + { + "path": "YFkyO6R1Frgw.wav", + "question": "Identify the occurrences of 'a large truck motor changing gears in the distance' along with their time frames.", + "choice_a": "[0.19, 0.51]", + "choice_b": "[7.27, 10.0]", + "choice_c": "[4.66, 5.75]", + "choice_d": "[7.34, 8.34]", + "answer_gt": "[7.27, 10.0]", + "answer_gt_word": "a large truck motor changing gears in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11703 + }, + { + "path": "YFm4oMQOP6Nc.wav", + "question": "List the periods when 'a large truck motor changing gears in the distance' is completely played in the sound.", + "choice_a": "[0.0, 1.14]", + "choice_b": "[7.27, 10.0]", + "choice_c": "[6.54, 6.72]", + "choice_d": "[7.38, 7.97]", + "answer_gt": "[7.27, 10.0]", + "answer_gt_word": "a large truck motor changing gears in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11704 + }, + { + "path": "YG9cpl27TYw8.wav", + "question": "How many seconds does each full presence of 'screeching' cover?", + "choice_a": "[0.11, 2.43]", + "choice_b": "[6.66, 7.99]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[6.67, 8.01]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "screeching", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11705 + }, + { + "path": "YGLD6PgwF7Mg.wav", + "question": "What is the range in seconds for each full 'screeching' in the recording?", + "choice_a": "[0.0, 4.81]", + "choice_b": "[5.63, 10.0]", + "choice_c": "[5.64, 6.39]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "screeching", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11706 + }, + { + "path": "YGdtMcEQmO4c.wav", + "question": "At what time intervals does 'another engine' occur in the sound file?", + "choice_a": "[0.0, 5.89]", + "choice_b": "[8.1, 10.0]", + "choice_c": "[8.1, 10.0]", + "choice_d": "[5.89, 8.12]", + "answer_gt": "[8.1, 10.0]", + "answer_gt_word": "another engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11707 + }, + { + "path": "YGoKR2cbKdxA.wav", + "question": "In what segments can 'another engine' be heard from start to finish?", + "choice_a": "[8.1, 10.0]", + "choice_b": "[3.17, 3.88]", + "choice_c": "[5.89, 6.38]", + "choice_d": "[0.74, 1.61]", + "answer_gt": "[8.1, 10.0]", + "answer_gt_word": "another engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11708 + }, + { + "path": "YHMoZBbpPBFk.wav", + "question": "In what time spans is 'a car grinds but fails to start' completely audible in this recording?", + "choice_a": "[2.45, 5.54]", + "choice_b": "[8.22, 9.89]", + "choice_c": "[4.13, 8.1]", + "choice_d": "[3.7, 6.13]", + "answer_gt": "[4.13, 8.1]", + "answer_gt_word": "a car grinds but fails to start", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11709 + }, + { + "path": "YHWBzPVfLFHg.wav", + "question": "What are the exact times when 'a dog panting' is present in the clip?", + "choice_a": "[0.19, 10.0]", + "choice_b": "[2.91, 3.79]", + "choice_c": "[4.33, 5.95]", + "choice_d": "[4.91, 9.5]", + "answer_gt": "[0.19, 10.0]", + "answer_gt_word": "a dog panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11710 + }, + { + "path": "YHhwFPF8BwSk.wav", + "question": "Where does 'a train engine hums nearby' appear from beginning to end in the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[5.69, 8.01]", + "choice_c": "[3.09, 6.55]", + "choice_d": "[4.09, 9.97]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a train engine hums nearby", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11711 + }, + { + "path": "YHmBBO3sV8DI.wav", + "question": "Where is 'a train horn blows' found in its entirety within the audio's timeline?", + "choice_a": "[0.29, 1.25]", + "choice_b": "[0.7, 3.59]", + "choice_c": "[1.58, 5.1]", + "choice_d": "[1.62, 7.88]", + "answer_gt": "[0.29, 1.25]", + "answer_gt_word": "a train horn blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11712 + }, + { + "path": "YHp_uadxXh4g.wav", + "question": "What are all the complete time segments of 'an adult female speaking' in this sound?", + "choice_a": "[0.14, 9.67]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[6.13, 9.04]", + "choice_d": "[9.79, 10.0]", + "answer_gt": "[0.14, 9.67]", + "answer_gt_word": "an adult female speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11713 + }, + { + "path": "YI-btJhP_ki8.wav", + "question": "At what timestamps does the complete 'an adult female speaking' occur?", + "choice_a": "[0.0, 0.69]", + "choice_b": "[0.14, 9.67]", + "choice_c": "[0.0, 0.31]", + "choice_d": "[0.89, 1.74]", + "answer_gt": "[0.14, 9.67]", + "answer_gt_word": "an adult female speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11714 + }, + { + "path": "YIPfaRF76gVU.wav", + "question": "Where do we hear 'a truck accelerates' in full within this sound recording?", + "choice_a": "[0.42, 1.17]", + "choice_b": "[8.28, 9.49]", + "choice_c": "[0.0, 9.48]", + "choice_d": "[8.28, 9.49]", + "answer_gt": "[0.0, 9.48]", + "answer_gt_word": "a truck accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11715 + }, + { + "path": "YIZffL8aQW9w.wav", + "question": "At what seconds do the full iterations of 'a horse neighs' start and end?", + "choice_a": "[0.94, 3.07]", + "choice_b": "[0.24, 1.88]", + "choice_c": "[3.05, 10.0]", + "choice_d": "[2.29, 6.08]", + "answer_gt": "[0.94, 3.07]", + "answer_gt_word": "a horse neighs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11716 + }, + { + "path": "YIhjS6l-BTio.wav", + "question": "Identify the occurrences of 'a speedboat accelerates loudly' along with their time frames.", + "choice_a": "[0.0, 9.41]", + "choice_b": "[5.55, 6.25]", + "choice_c": "[5.2, 9.33]", + "choice_d": "[5.42, 9.56]", + "answer_gt": "[0.0, 9.41]", + "answer_gt_word": "a speedboat accelerates loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11717 + }, + { + "path": "YJC8TDJnubFg.wav", + "question": "At which time markers does 'running occurs' fully unfold in the sound?", + "choice_a": "[3.15, 8.38]", + "choice_b": "[0.11, 1.48]", + "choice_c": "[2.15, 8.71]", + "choice_d": "[2.82, 4.9]", + "answer_gt": "[3.15, 8.38]", + "answer_gt_word": "running occurs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11718 + }, + { + "path": "YJFk4Qyn58CY.wav", + "question": "What are the intervals in seconds of each 'a large engine goes by' appearance?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[6.98, 9.54]", + "choice_c": "[4.76, 6.96]", + "choice_d": "[3.67, 8.22]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a large engine goes by", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11719 + }, + { + "path": "YJP_czEZIOb4.wav", + "question": "What second marks correspond to the entire span of 'a train is sounding its horn'?", + "choice_a": "[0.75, 1.9]", + "choice_b": "[0.72, 4.74]", + "choice_c": "[1.0, 9.15]", + "choice_d": "[3.17, 9.88]", + "answer_gt": "[0.75, 1.9]", + "answer_gt_word": "a train is sounding its horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11720 + }, + { + "path": "YJcT3vhg_Ie8.wav", + "question": "What are the intervals in seconds of each 'a large engine idling' appearance?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.16, 6.4]", + "choice_d": "[6.28, 9.8]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a large engine idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11721 + }, + { + "path": "YJiPwWia5ryk.wav", + "question": "Identify the periods when 'the screech of wheels on the rails' fully resonates in the sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.83, 7.93]", + "choice_c": "[4.28, 8.31]", + "choice_d": "[6.36, 9.93]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "the screech of wheels on the rails", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11722 + }, + { + "path": "YKVbmN9ZRg5Q.wav", + "question": "How is 'a loud continuous horn' distributed over the sound's duration?", + "choice_a": "[0.75, 9.99]", + "choice_b": "[0.0, 0.78]", + "choice_c": "[2.68, 7.87]", + "choice_d": "[2.79, 9.45]", + "answer_gt": "[0.75, 9.99]", + "answer_gt_word": "a loud continuous horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11723 + }, + { + "path": "YKq0Dbp3C4d0.wav", + "question": "Where does 'a horn is triggered far away' appear from beginning to end in the sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 2.28]", + "choice_c": "[3.45, 9.52]", + "choice_d": "[3.33, 9.13]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a horn is triggered far away", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11724 + }, + { + "path": "YLBixNsRsbpE.wav", + "question": "What timestamps mark the full occurrences of 'chickens and roosters clucking'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[8.86, 10.0]", + "choice_c": "[0.65, 1.84]", + "choice_d": "[0.66, 7.88]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "chickens and roosters clucking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11725 + }, + { + "path": "YLlffd6cn5jY.wav", + "question": "List the periods when 'an running engine revving' is completely played in the sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[2.14, 6.88]", + "choice_d": "[6.69, 7.23]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an running engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11726 + }, + { + "path": "YMM9O504q7Zc.wav", + "question": "What are the specific start and stop times of 'an running engine revving' in the sound?", + "choice_a": "[0.15, 1.05]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[2.91, 3.5]", + "choice_d": "[4.63, 5.15]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an running engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11727 + }, + { + "path": "YMlb1RW4-LRg.wav", + "question": "Enumerate the second-based occurrences of 'an running engine revving' in full.", + "choice_a": "[0.22, 2.46]", + "choice_b": "[3.27, 4.97]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[1.68, 2.42]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an running engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11728 + }, + { + "path": "YMtmE6qBxURA.wav", + "question": "What timestamps mark the full occurrences of 'a water vehicle travels through the water'?", + "choice_a": "[0.0, 9.84]", + "choice_b": "[2.8, 3.75]", + "choice_c": "[4.82, 5.59]", + "choice_d": "[0.0, 9.84]", + "answer_gt": "[0.0, 9.84]", + "answer_gt_word": "a water vehicle travels through the water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11729 + }, + { + "path": "YMvHpNzDpC6Q.wav", + "question": "What ranges in seconds capture 'the other chuckles' in full?", + "choice_a": "[0.32, 1.46]", + "choice_b": "[2.59, 2.99]", + "choice_c": "[3.97, 4.5]", + "choice_d": "[5.91, 6.52]", + "answer_gt": "[5.91, 6.52]", + "answer_gt_word": "the other chuckles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11730 + }, + { + "path": "YN20hpbx31NM.wav", + "question": "What are the distinct time ranges of 'a truck engine running idle' within the audio?", + "choice_a": "[0.0, 9.04]", + "choice_b": "[8.6, 8.84]", + "choice_c": "[5.88, 6.65]", + "choice_d": "[5.46, 6.0]", + "answer_gt": "[0.0, 9.04]", + "answer_gt_word": "a truck engine running idle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11731 + }, + { + "path": "YNUCh9Xtbjkw.wav", + "question": "Where do we hear 'a vehicle engine runs' in full within this sound recording?", + "choice_a": "[0.01, 5.72]", + "choice_b": "[5.67, 6.8]", + "choice_c": "[0.79, 2.72]", + "choice_d": "[1.76, 7.84]", + "answer_gt": "[0.01, 5.72]", + "answer_gt_word": "a vehicle engine runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11732 + }, + { + "path": "YNek57Uq--bw.wav", + "question": "Locate every segment where 'rumbling thunder' is completely audible.", + "choice_a": "[3.4, 5.16]", + "choice_b": "[7.15, 8.15]", + "choice_c": "[8.98, 9.53]", + "choice_d": "[5.91, 9.99]", + "answer_gt": "[5.91, 9.99]", + "answer_gt_word": "rumbling thunder", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11733 + }, + { + "path": "YNfGOKIniyyE.wav", + "question": "What are the intervals in seconds of each 'revving of an engine' appearance?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.0, 4.2]", + "choice_d": "[7.05, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "revving of an engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11734 + }, + { + "path": "YNnXucrfqnJ8.wav", + "question": "What time segments showcase 'splashing water' in its entirety?", + "choice_a": "[0.01, 10.01]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.84, 6.78]", + "choice_d": "[6.73, 7.69]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "splashing water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11735 + }, + { + "path": "YO5EGxn49f5w.wav", + "question": "What time segments delineate the presence of 'splashing water'?", + "choice_a": "[1.04, 1.86]", + "choice_b": "[6.36, 9.99]", + "choice_c": "[1.06, 1.86]", + "choice_d": "[0.01, 10.01]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "splashing water", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11736 + }, + { + "path": "YO9xYzACrBjA.wav", + "question": "How are the complete segments of 'an airplane propeller spins loudly' timed in this audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[8.38, 9.73]", + "choice_c": "[2.71, 5.08]", + "choice_d": "[2.88, 7.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an airplane propeller spins loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11737 + }, + { + "path": "YOVS0Z30_mMU.wav", + "question": "Where do we hear 'humming' in full within this sound recording?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.02, 10.0]", + "choice_c": "[2.88, 7.06]", + "choice_d": "[9.59, 10.0]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11738 + }, + { + "path": "YOpmtLXu5NJI.wav", + "question": "What are all the complete time segments of 'a loud train horn sounding' in this sound?", + "choice_a": "[0.01, 1.31]", + "choice_b": "[4.45, 6.65]", + "choice_c": "[9.35, 10.0]", + "choice_d": "[4.58, 10.0]", + "answer_gt": "[0.01, 1.31]", + "answer_gt_word": "a loud train horn sounding", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11739 + }, + { + "path": "YOt0bN_hz2ec.wav", + "question": "How long does 'a honk' last each time it appears in the audio?", + "choice_a": "[0.77, 2.44]", + "choice_b": "[0.0, 10.01]", + "choice_c": "[0.43, 4.33]", + "choice_d": "[2.02, 9.0]", + "answer_gt": "[0.77, 2.44]", + "answer_gt_word": "a honk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11740 + }, + { + "path": "YPSdbflW9mQ0.wav", + "question": "What times mark the complete cycles of 'humming' in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.47, 2.3]", + "choice_c": "[6.29, 8.77]", + "choice_d": "[0.08, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11741 + }, + { + "path": "YPY_wYey3P14.wav", + "question": "What are the full time segments when 'a race car engine revving around a track' appears in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.88, 2.57]", + "choice_c": "[0.95, 4.42]", + "choice_d": "[2.44, 4.73]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a race car engine revving around a track", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11742 + }, + { + "path": "YQWVZ3oMc4g0.wav", + "question": "Where is 'a car skids' consistently heard throughout the sound?", + "choice_a": "[3.16, 4.29]", + "choice_b": "[2.92, 9.08]", + "choice_c": "[2.22, 7.7]", + "choice_d": "[4.42, 9.44]", + "answer_gt": "[3.16, 4.29]", + "answer_gt_word": "a car skids", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11743 + }, + { + "path": "YQYQvPofLYOA.wav", + "question": "Define the time intervals that fully contain 'a car skids' in this sound.", + "choice_a": "[0.28, 0.76]", + "choice_b": "[3.16, 4.29]", + "choice_c": "[3.74, 4.3]", + "choice_d": "[0.92, 1.43]", + "answer_gt": "[3.16, 4.29]", + "answer_gt_word": "a car skids", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11744 + }, + { + "path": "YRdHDbFchcV8.wav", + "question": "How many seconds does each full presence of 'an emergency vehicle sounds siren' cover?", + "choice_a": "[0.0, 7.21]", + "choice_b": "[5.42, 7.21]", + "choice_c": "[5.35, 7.21]", + "choice_d": "[1.55, 6.24]", + "answer_gt": "[0.0, 7.21]", + "answer_gt_word": "an emergency vehicle sounds siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11745 + }, + { + "path": "YS3dsryb4eos.wav", + "question": "Identify the occurrences of 'a horn is triggered' along with their time frames.", + "choice_a": "[1.27, 3.01]", + "choice_b": "[2.36, 4.68]", + "choice_c": "[3.45, 5.2]", + "choice_d": "[4.24, 7.79]", + "answer_gt": "[1.27, 3.01]", + "answer_gt_word": "a horn is triggered", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11746 + }, + { + "path": "YSEHB4eYLLPc.wav", + "question": "In what segments can 'a horn is triggered' be heard from start to finish?", + "choice_a": "[0.14, 4.35]", + "choice_b": "[1.29, 4.34]", + "choice_c": "[2.12, 7.35]", + "choice_d": "[4.32, 9.75]", + "answer_gt": "[0.14, 4.35]", + "answer_gt_word": "a horn is triggered", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11747 + }, + { + "path": "YSEql80jbheU.wav", + "question": "Provide the start and end times of 'an engine works in idle' within this audio.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.18, 6.93]", + "choice_c": "[3.62, 7.82]", + "choice_d": "[6.91, 8.19]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine works in idle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11748 + }, + { + "path": "YSNIaYhri76w.wav", + "question": "At what timestamps does the complete 'a pig squeals' occur?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.25, 2.53]", + "choice_c": "[6.56, 6.83]", + "choice_d": "[7.97, 9.35]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a pig squeals", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11749 + }, + { + "path": "YSNIaYhri76w.wav", + "question": "Enumerate the second-based occurrences of 'a pig oinks' in full.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[7.97, 9.32]", + "choice_c": "[1.44, 7.84]", + "choice_d": "[5.51, 9.94]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a pig oinks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11750 + }, + { + "path": "YSNIaYhri76w.wav", + "question": "What are the comprehensive segments of 'man speaking' in time?", + "choice_a": "[2.28, 2.55]", + "choice_b": "[6.48, 6.85]", + "choice_c": "[7.98, 9.34]", + "choice_d": "[2.65, 6.83]", + "answer_gt": "[7.98, 9.34]", + "answer_gt_word": "man speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11751 + }, + { + "path": "YSiMabSrCdJw.wav", + "question": "What are the full time segments when 'a train whistle blows' appears in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.26, 3.83]", + "choice_d": "[2.7, 7.37]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a train whistle blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11752 + }, + { + "path": "YSmB8tjUGVNE.wav", + "question": "Define the time intervals that fully contain 'rustling' in this sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[2.01, 3.7]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11753 + }, + { + "path": "YUsncYzpYSk8.wav", + "question": "Identify the complete instances of 'two dogs bark' within the sound timeline.", + "choice_a": "[0.78, 10.0]", + "choice_b": "[0.01, 10.0]", + "choice_c": "[0.43, 3.38]", + "choice_d": "[2.99, 3.92]", + "answer_gt": "[0.78, 10.0]", + "answer_gt_word": "two dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11754 + }, + { + "path": "YUwQhrCDchio.wav", + "question": "In what time spans is 'two dogs bark' completely audible in this recording?", + "choice_a": "[0.59, 5.92]", + "choice_b": "[0.78, 10.0]", + "choice_c": "[6.49, 6.95]", + "choice_d": "[9.11, 9.73]", + "answer_gt": "[0.78, 10.0]", + "answer_gt_word": "two dogs bark", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11755 + }, + { + "path": "YVOXl8iR-HnI.wav", + "question": "At what time points is 'humming of a distant jet engine passing' fully present in the recording?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.05, 3.24]", + "choice_c": "[0.53, 4.09]", + "choice_d": "[2.35, 6.39]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "humming of a distant jet engine passing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11756 + }, + { + "path": "YVc5wX72RjN0.wav", + "question": "Capture the timeline where 'laughter' is fully included in the sound.", + "choice_a": "[0.0, 1.24]", + "choice_b": "[9.6, 9.99]", + "choice_c": "[3.9, 5.87]", + "choice_d": "[6.73, 10.0]", + "answer_gt": "[9.6, 9.99]", + "answer_gt_word": "laughter", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11757 + }, + { + "path": "YVhIYEoQseoM.wav", + "question": "How many seconds does each full presence of 'more barking' cover?", + "choice_a": "[0.0, 0.8]", + "choice_b": "[3.58, 3.93]", + "choice_c": "[7.34, 8.18]", + "choice_d": "[7.35, 8.08]", + "answer_gt": "[7.35, 8.08]", + "answer_gt_word": "more barking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11758 + }, + { + "path": "YVmSHONdqvng.wav", + "question": "How long in seconds does 'racecars revs' last from start to end?", + "choice_a": "[0.01, 10.0]", + "choice_b": "[4.78, 8.8]", + "choice_c": "[9.7, 10.0]", + "choice_d": "[2.73, 6.75]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "racecars revs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11759 + }, + { + "path": "YW-0MMPGAD7k.wav", + "question": "Identify the complete instances of 'a man talks' within the sound timeline.", + "choice_a": "[0.46, 2.75]", + "choice_b": "[6.74, 7.46]", + "choice_c": "[2.22, 10.0]", + "choice_d": "[1.72, 4.37]", + "answer_gt": "[0.46, 2.75]", + "answer_gt_word": "a man talks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11760 + }, + { + "path": "YW7OJevEgq7w.wav", + "question": "Where does 'panting' appear from beginning to end in the sound?", + "choice_a": "[1.31, 4.8]", + "choice_b": "[6.26, 8.06]", + "choice_c": "[9.31, 10.0]", + "choice_d": "[1.32, 10.0]", + "answer_gt": "[1.32, 10.0]", + "answer_gt_word": "panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11761 + }, + { + "path": "YWMwYUbpKzfY.wav", + "question": "How is 'panting' distributed over the sound's duration?", + "choice_a": "[0.96, 1.29]", + "choice_b": "[6.99, 7.43]", + "choice_c": "[1.32, 10.0]", + "choice_d": "[6.26, 6.42]", + "answer_gt": "[1.32, 10.0]", + "answer_gt_word": "panting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11762 + }, + { + "path": "YXTGcypOsLgw.wav", + "question": "Enumerate the second-based occurrences of 'faint murmuring' in full.", + "choice_a": "[2.79, 3.99]", + "choice_b": "[6.15, 10.0]", + "choice_c": "[0.0, 6.17]", + "choice_d": "[2.28, 7.1]", + "answer_gt": "[0.0, 6.17]", + "answer_gt_word": "faint murmuring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11763 + }, + { + "path": "YXVlnx4ofu2k.wav", + "question": "At what timestamps does the complete 'faint murmuring' occur?", + "choice_a": "[0.0, 5.68]", + "choice_b": "[8.12, 10.0]", + "choice_c": "[0.0, 5.68]", + "choice_d": "[0.0, 6.17]", + "answer_gt": "[0.0, 6.17]", + "answer_gt_word": "faint murmuring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11764 + }, + { + "path": "YXm0LegZNUOQ.wav", + "question": "List the specific seconds of 'a train horn blows in the distance' appearances.", + "choice_a": "[0.0, 3.21]", + "choice_b": "[2.53, 4.16]", + "choice_c": "[3.16, 5.45]", + "choice_d": "[3.61, 9.98]", + "answer_gt": "[0.0, 3.21]", + "answer_gt_word": "a train horn blows in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11765 + }, + { + "path": "YXn4SruTQFfY.wav", + "question": "Identify the periods when 'people talk' fully resonates in the sound.", + "choice_a": "[0.0, 6.61]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.81, 7.74]", + "choice_d": "[7.16, 9.99]", + "answer_gt": "[0.0, 6.61]", + "answer_gt_word": "people talk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11766 + }, + { + "path": "YYZOEAWhD7js.wav", + "question": "Capture the timeline where 'a train moves closer' is fully included in the sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[3.11, 3.55]", + "choice_c": "[2.2, 8.4]", + "choice_d": "[2.97, 9.74]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a train moves closer", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11767 + }, + { + "path": "YYp6CI7K4UcY.wav", + "question": "At what time intervals does 'a train moves closer' occur in the sound file?", + "choice_a": "[0.0, 3.16]", + "choice_b": "[4.85, 10.02]", + "choice_c": "[0.95, 1.33]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a train moves closer", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11768 + }, + { + "path": "YZLHtRWod3wM.wav", + "question": "What time segments showcase 'idles loudly' in its entirety?", + "choice_a": "[3.61, 10.01]", + "choice_b": "[0.5, 6.62]", + "choice_c": "[2.64, 6.77]", + "choice_d": "[2.74, 8.27]", + "answer_gt": "[3.61, 10.01]", + "answer_gt_word": "idles loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11769 + }, + { + "path": "YZVlZro7vrxw.wav", + "question": "At what time intervals does 'cry' occur in the sound file?", + "choice_a": "[5.95, 7.98]", + "choice_b": "[3.61, 4.62]", + "choice_c": "[5.54, 6.13]", + "choice_d": "[8.07, 9.16]", + "answer_gt": "[5.95, 7.98]", + "answer_gt_word": "cry", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11770 + }, + { + "path": "YZq8UkUAhLZQ.wav", + "question": "What time segments delineate the presence of 'an engine idles'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.0, 10.02]", + "choice_c": "[0.65, 1.07]", + "choice_d": "[5.09, 5.97]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11771 + }, + { + "path": "Y_GI7meqlYZk.wav", + "question": "How are the complete segments of 'an engine idles' timed in this audio?", + "choice_a": "[0.17, 1.7]", + "choice_b": "[4.0, 4.74]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[8.26, 9.6]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an engine idles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11772 + }, + { + "path": "YaByChd13yEA.wav", + "question": "Identify the complete instances of 'an engine revs the whole time' within the sound timeline.", + "choice_a": "[0.0, 4.38]", + "choice_b": "[5.49, 8.39]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[0.0, 8.41]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine revs the whole time", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11773 + }, + { + "path": "YaC_Ynuy_cl0.wav", + "question": "Locate every segment where 'a small engine runs' is completely audible.", + "choice_a": "[0.0, 10.01]", + "choice_b": "[2.09, 6.75]", + "choice_c": "[7.42, 9.99]", + "choice_d": "[2.1, 6.61]", + "answer_gt": "[0.0, 10.01]", + "answer_gt_word": "a small engine runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11774 + }, + { + "path": "YarFwKX6iU38.wav", + "question": "Which sections of the sound contain 'a high pitched engine running continuously' in their entirety?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[3.42, 4.66]", + "choice_c": "[1.34, 7.89]", + "choice_d": "[5.26, 9.3]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a high pitched engine running continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11775 + }, + { + "path": "Yb4xHgy3Fv3g.wav", + "question": "Identify the periods when 'a high pitched engine running continuously' fully resonates in the sound.", + "choice_a": "[0.0, 1.2]", + "choice_b": "[3.36, 10.0]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[8.05, 9.01]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a high pitched engine running continuously", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11776 + }, + { + "path": "YbPOo7UBbwOU.wav", + "question": "How long does 'rustles' last each time it appears in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.95, 3.68]", + "choice_c": "[0.33, 1.45]", + "choice_d": "[4.41, 4.67]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rustles", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11777 + }, + { + "path": "YbWYqngojj7Y.wav", + "question": "How long in seconds does 'an engine revving' last from start to end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.23, 3.45]", + "choice_c": "[2.89, 4.49]", + "choice_d": "[3.66, 7.48]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine revving", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11778 + }, + { + "path": "Ybpv_LneHmfU.wav", + "question": "Enumerate the second-based occurrences of 'humming of a nearby jet engine' in full.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.35, 8.02]", + "choice_c": "[0.84, 8.21]", + "choice_d": "[3.03, 9.04]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "humming of a nearby jet engine", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11779 + }, + { + "path": "YcE7sLRthlZM.wav", + "question": "At what timestamps does the complete 'a fire truck siren wails' occur?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.34, 4.66]", + "choice_c": "[1.51, 4.94]", + "choice_d": "[3.92, 7.77]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a fire truck siren wails", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11780 + }, + { + "path": "YcLtItZ0WGQc.wav", + "question": "Document the full appearances of 'rain falling and pattering on a cloth surface' in terms of seconds.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[8.12, 9.56]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rain falling and pattering on a cloth surface", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11781 + }, + { + "path": "YcO4_J-DNZvM.wav", + "question": "Where do we hear 'race car engine accelerates' in full within this sound recording?", + "choice_a": "[0.74, 3.3]", + "choice_b": "[3.88, 4.64]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[0.9, 7.36]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "race car engine accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11782 + }, + { + "path": "YcSURZuti4dU.wav", + "question": "What time segments delineate the presence of 'rustling'?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[2.95, 3.51]", + "choice_c": "[4.1, 7.65]", + "choice_d": "[8.88, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11783 + }, + { + "path": "Yd55UQEWwDN8.wav", + "question": "List the periods when 'a fast car accelerates' is completely played in the sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.28, 6.7]", + "choice_c": "[1.58, 8.17]", + "choice_d": "[3.0, 9.2]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a fast car accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11784 + }, + { + "path": "YdKCJpCevfCI.wav", + "question": "At which seconds does 'someone speaks' fully register in the sound?", + "choice_a": "[2.12, 3.25]", + "choice_b": "[3.8, 5.38]", + "choice_c": "[6.47, 8.41]", + "choice_d": "[5.94, 5.95]", + "answer_gt": "[6.47, 8.41]", + "answer_gt_word": "someone speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11785 + }, + { + "path": "YdgnWKrBl4OQ.wav", + "question": "Where is 'rustling' consistently heard throughout the sound?", + "choice_a": "[1.02, 1.56]", + "choice_b": "[4.82, 5.72]", + "choice_c": "[6.84, 7.14]", + "choice_d": "[0.0, 8.69]", + "answer_gt": "[0.0, 8.69]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11786 + }, + { + "path": "Ye30uwJvj4Io.wav", + "question": "What are the intervals in seconds of each 'rustling' appearance?", + "choice_a": "[0.0, 5.38]", + "choice_b": "[7.4, 8.06]", + "choice_c": "[0.0, 8.69]", + "choice_d": "[5.61, 6.66]", + "answer_gt": "[0.0, 8.69]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11787 + }, + { + "path": "Ye6y_ejwxrng.wav", + "question": "At what timestamps does the complete 'a soft fart' occur?", + "choice_a": "[0.0, 1.51]", + "choice_b": "[6.48, 7.56]", + "choice_c": "[9.03, 10.0]", + "choice_d": "[2.26, 2.78]", + "answer_gt": "[2.26, 2.78]", + "answer_gt_word": "a soft fart", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11788 + }, + { + "path": "YeAroMk9QxuY.wav", + "question": "List the periods when 'vehicles honk very loudly' is completely played in the sound.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[8.53, 10.0]", + "choice_c": "[2.11, 3.57]", + "choice_d": "[3.48, 5.99]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "vehicles honk very loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11789 + }, + { + "path": "Yet5VaXFQKzs.wav", + "question": "Determine the full durations of 'a train whistle blowing' in the audio file.", + "choice_a": "[9.38, 9.63]", + "choice_b": "[3.5, 7.27]", + "choice_c": "[3.98, 8.27]", + "choice_d": "[7.0, 9.49]", + "answer_gt": "[9.38, 9.63]", + "answer_gt_word": "a train whistle blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11790 + }, + { + "path": "YfKwl4gOlPE4.wav", + "question": "What are the comprehensive timeframes of 'rain falls' in this clip?", + "choice_a": "[0.66, 4.54]", + "choice_b": "[8.71, 10.0]", + "choice_c": "[0.0, 8.17]", + "choice_d": "[0.0, 0.38]", + "answer_gt": "[0.0, 8.17]", + "answer_gt_word": "rain falls", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11791 + }, + { + "path": "Yg26ODac1y3o.wav", + "question": "Enumerate the second-based occurrences of 'a man talks in the background' in full.", + "choice_a": "[0.17, 0.71]", + "choice_b": "[2.73, 5.7]", + "choice_c": "[6.13, 6.43]", + "choice_d": "[8.76, 9.4]", + "answer_gt": "[0.17, 0.71]", + "answer_gt_word": "a man talks in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11792 + }, + { + "path": "YgA6n2-9zxKM.wav", + "question": "Capture the timeline where 'a motor vehicle engine is idling' is fully included in the sound.", + "choice_a": "[0.0, 9.99]", + "choice_b": "[2.39, 5.71]", + "choice_c": "[7.03, 7.61]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 9.99]", + "answer_gt_word": "a motor vehicle engine is idling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11793 + }, + { + "path": "YgJ_qu2EC8Kw.wav", + "question": "How long in seconds does 'a train horn blowing twice' last from start to end?", + "choice_a": "[5.74, 7.65]", + "choice_b": "[3.12, 3.75]", + "choice_c": "[1.78, 5.29]", + "choice_d": "[2.2, 6.93]", + "answer_gt": "[5.74, 7.65]", + "answer_gt_word": "a train horn blowing twice", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11794 + }, + { + "path": "YgWVY03cqNTA.wav", + "question": "What are the boundaries in seconds for each 'motor running' presence?", + "choice_a": "[0.0, 5.55]", + "choice_b": "[6.47, 8.21]", + "choice_c": "[0.43, 2.28]", + "choice_d": "[2.06, 3.62]", + "answer_gt": "[0.0, 5.55]", + "answer_gt_word": "motor running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11795 + }, + { + "path": "YgXU3IbRXYIQ.wav", + "question": "At what seconds do the full iterations of 'dull roar' start and end?", + "choice_a": "[0.01, 9.99]", + "choice_b": "[6.83, 7.33]", + "choice_c": "[8.28, 10.0]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.01, 9.99]", + "answer_gt_word": "dull roar", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11796 + }, + { + "path": "YgZAnkBJMK5k.wav", + "question": "At what seconds do the full iterations of 'music plays in the background' start and end?", + "choice_a": "[0.0, 5.44]", + "choice_b": "[6.25, 6.44]", + "choice_c": "[7.62, 8.5]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "music plays in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11797 + }, + { + "path": "YglyZUCN3BfU.wav", + "question": "In what time spans is 'a vehicle engine starting up' completely audible in this recording?", + "choice_a": "[0.85, 1.63]", + "choice_b": "[0.86, 10.0]", + "choice_c": "[1.72, 10.0]", + "choice_d": "[8.53, 8.93]", + "answer_gt": "[0.85, 1.63]", + "answer_gt_word": "a vehicle engine starting up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11798 + }, + { + "path": "YgqYEFSj6vBA.wav", + "question": "Document the full appearances of 'race car accelerating' in terms of seconds.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.4, 6.37]", + "choice_c": "[2.17, 6.9]", + "choice_d": "[5.31, 7.42]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "race car accelerating", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11799 + }, + { + "path": "YhFHkJ2qWm90.wav", + "question": "What second marks correspond to the entire span of 'a car honks'?", + "choice_a": "[6.77, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[5.88, 9.08]", + "choice_d": "[6.46, 9.2]", + "answer_gt": "[6.77, 10.0]", + "answer_gt_word": "a car honks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11800 + }, + { + "path": "YhogpqdigubM.wav", + "question": "What ranges in seconds capture 'a train horn blowing' in full?", + "choice_a": "[0.19, 3.19]", + "choice_b": "[4.1, 8.77]", + "choice_c": "[8.74, 10.0]", + "choice_d": "[1.44, 9.7]", + "answer_gt": "[8.74, 10.0]", + "answer_gt_word": "a train horn blowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11801 + }, + { + "path": "Yimbp68i3pxo.wav", + "question": "At which seconds does 'humming' fully register in the sound?", + "choice_a": "[0.0, 9.31]", + "choice_b": "[0.24, 0.57]", + "choice_c": "[1.38, 1.7]", + "choice_d": "[9.25, 10.0]", + "answer_gt": "[0.0, 9.31]", + "answer_gt_word": "humming", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11802 + }, + { + "path": "YjCExyUac_xM.wav", + "question": "How many seconds does each full presence of 'an engine buzzing' cover?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.82, 3.55]", + "choice_c": "[4.31, 5.53]", + "choice_d": "[6.02, 9.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine buzzing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11803 + }, + { + "path": "YjHv726cD6TQ.wav", + "question": "Which sections of the sound contain 'a man speaks' in their entirety?", + "choice_a": "[5.47, 8.51]", + "choice_b": "[2.24, 3.54]", + "choice_c": "[4.13, 10.0]", + "choice_d": "[5.74, 8.02]", + "answer_gt": "[5.47, 8.51]", + "answer_gt_word": "a man speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11804 + }, + { + "path": "YjOgB2_9y9S8.wav", + "question": "Identify the periods when 'a loud horn' fully resonates in the sound.", + "choice_a": "[2.37, 4.07]", + "choice_b": "[7.81, 8.46]", + "choice_c": "[9.45, 10.0]", + "choice_d": "[4.55, 7.31]", + "answer_gt": "[4.55, 7.31]", + "answer_gt_word": "a loud horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11805 + }, + { + "path": "YjbIEjeGUpMc.wav", + "question": "What are the exact times when 'someone is speaking over a speaker system' is present in the clip?", + "choice_a": "[0.11, 3.37]", + "choice_b": "[2.53, 3.13]", + "choice_c": "[2.99, 7.55]", + "choice_d": "[5.21, 9.86]", + "answer_gt": "[0.11, 3.37]", + "answer_gt_word": "someone is speaking over a speaker system", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11806 + }, + { + "path": "YkC2t97-pw5Q.wav", + "question": "At what points does 'someone is speaking over a speaker system' exist completely in the timeline?", + "choice_a": "[0.11, 3.37]", + "choice_b": "[4.8, 6.28]", + "choice_c": "[8.37, 9.94]", + "choice_d": "[1.7, 4.77]", + "answer_gt": "[0.11, 3.37]", + "answer_gt_word": "someone is speaking over a speaker system", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11807 + }, + { + "path": "YkFyDYvCvOqo.wav", + "question": "What timestamps mark the full occurrences of 'cars rev their engines in the distance'?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[2.18, 8.02]", + "choice_c": "[2.53, 8.3]", + "choice_d": "[6.47, 8.94]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "cars rev their engines in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11808 + }, + { + "path": "YlTyPKec30Bk.wav", + "question": "At what timestamps does the complete 'motor vehicles roar by in the distance' occur?", + "choice_a": "[0.01, 10.01]", + "choice_b": "[0.46, 7.14]", + "choice_c": "[2.72, 8.11]", + "choice_d": "[5.34, 9.51]", + "answer_gt": "[0.01, 10.01]", + "answer_gt_word": "motor vehicles roar by in the distance", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11809 + }, + { + "path": "YlX3k5p2I_g0.wav", + "question": "At what time points is 'another man shouting' fully present in the recording?", + "choice_a": "[0.27, 1.42]", + "choice_b": "[2.2, 3.37]", + "choice_c": "[2.79, 3.4]", + "choice_d": "[5.0, 10.02]", + "answer_gt": "[2.79, 3.4]", + "answer_gt_word": "another man shouting", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11810 + }, + { + "path": "YmHT7IM34WTY.wav", + "question": "How many seconds does 'sloshing water running' span in each occurrence?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.66, 5.55]", + "choice_d": "[2.63, 8.38]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sloshing water running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11811 + }, + { + "path": "YmvbjcYC2TaY.wav", + "question": "What are the comprehensive segments of 'sloshing water running' in time?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[9.61, 10.0]", + "choice_c": "[0.0, 2.78]", + "choice_d": "[5.57, 9.47]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sloshing water running", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11812 + }, + { + "path": "Yn7CbU6TM7Lw.wav", + "question": "List the specific seconds of 'clanking' appearances.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.36, 2.28]", + "choice_c": "[3.82, 4.99]", + "choice_d": "[0.5, 2.89]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "clanking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11813 + }, + { + "path": "YnKQ6BTfiAR8.wav", + "question": "What ranges in seconds capture 'a horn is triggered' in full?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.62, 4.11]", + "choice_c": "[2.07, 5.21]", + "choice_d": "[2.39, 9.17]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a horn is triggered", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11814 + }, + { + "path": "YnQy5YuL-W6A.wav", + "question": "What times mark the complete cycles of 'continuous rain drops' in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[1.66, 4.47]", + "choice_c": "[0.79, 4.98]", + "choice_d": "[2.69, 9.76]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "continuous rain drops", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11815 + }, + { + "path": "YnWEbiJc6zxc.wav", + "question": "At what timestamps does the complete 'another woman speaks loudly further away' occur?", + "choice_a": "[0.08, 0.74]", + "choice_b": "[2.46, 2.71]", + "choice_c": "[4.22, 4.84]", + "choice_d": "[5.55, 8.81]", + "answer_gt": "[5.55, 8.81]", + "answer_gt_word": "another woman speaks loudly further away", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11816 + }, + { + "path": "YnriEa916FaU.wav", + "question": "What times mark the complete cycles of 'a dog whimpering' in the audio?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.15, 0.43]", + "choice_d": "[1.46, 4.43]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a dog whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11817 + }, + { + "path": "Yo7VIBi-2wg4.wav", + "question": "What time segments delineate the presence of 'a vehicle engine accelerates'?", + "choice_a": "[2.05, 6.0]", + "choice_b": "[1.0, 7.95]", + "choice_c": "[2.47, 8.56]", + "choice_d": "[4.53, 9.29]", + "answer_gt": "[2.05, 6.0]", + "answer_gt_word": "a vehicle engine accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11818 + }, + { + "path": "YoHfoEcqtWgs.wav", + "question": "Identify the occurrences of 'water runs' along with their time frames.", + "choice_a": "[0.0, 0.35]", + "choice_b": "[2.69, 10.0]", + "choice_c": "[3.42, 4.01]", + "choice_d": "[5.32, 6.59]", + "answer_gt": "[2.69, 10.0]", + "answer_gt_word": "water runs", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11819 + }, + { + "path": "YoN0IcZaHD_8.wav", + "question": "What time segments delineate the presence of 'a tools drilling wood'?", + "choice_a": "[0.0, 0.61]", + "choice_b": "[8.57, 10.0]", + "choice_c": "[1.22, 5.23]", + "choice_d": "[7.04, 8.32]", + "answer_gt": "[1.22, 5.23]", + "answer_gt_word": "a tools drilling wood", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11820 + }, + { + "path": "YodJmGjFIhX8.wav", + "question": "Provide the start and end times of 'city sounds hail in the background' within this audio.", + "choice_a": "[0.0, 10.02]", + "choice_b": "[5.5, 10.0]", + "choice_c": "[1.98, 2.89]", + "choice_d": "[3.64, 5.23]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "city sounds hail in the background", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11821 + }, + { + "path": "YoqD-U-HYM6I.wav", + "question": "What are the full time segments when 'a woman speaking' appears in this sound?", + "choice_a": "[0.0, 2.01]", + "choice_b": "[4.63, 9.7]", + "choice_c": "[2.68, 4.11]", + "choice_d": "[7.06, 9.12]", + "answer_gt": "[2.68, 4.11]", + "answer_gt_word": "a woman speaking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11822 + }, + { + "path": "YpWQeV08kYR0.wav", + "question": "How long in seconds does 'an ambulance sounds the siren' last from start to end?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.34, 7.96]", + "choice_c": "[1.41, 8.4]", + "choice_d": "[4.99, 8.47]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an ambulance sounds the siren", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11823 + }, + { + "path": "YpgzTXMgU1FM.wav", + "question": "What are all the complete time segments of 'a train goes by fast blowing a horn' in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 5.85]", + "choice_c": "[3.05, 8.11]", + "choice_d": "[7.44, 9.96]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a train goes by fast blowing a horn", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11824 + }, + { + "path": "Ypo-iqaVg9z8.wav", + "question": "How many seconds does each full presence of 'a dog whimpering' cover?", + "choice_a": "[0.1, 10.0]", + "choice_b": "[0.1, 9.98]", + "choice_c": "[0.86, 3.64]", + "choice_d": "[5.39, 5.66]", + "answer_gt": "[0.1, 10.0]", + "answer_gt_word": "a dog whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11825 + }, + { + "path": "YptIksg9KEac.wav", + "question": "Enumerate the second-based occurrences of 'a person clapping' in full.", + "choice_a": "[1.79, 2.3]", + "choice_b": "[7.54, 10.0]", + "choice_c": "[4.5, 4.79]", + "choice_d": "[4.88, 9.0]", + "answer_gt": "[1.79, 2.3]", + "answer_gt_word": "a person clapping", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11826 + }, + { + "path": "YqGJKXoKTN4A.wav", + "question": "What are the comprehensive segments of 'a stream of water gently flowing' in time?", + "choice_a": "[0.09, 10.02]", + "choice_b": "[0.07, 2.57]", + "choice_c": "[4.04, 4.55]", + "choice_d": "[5.55, 10.0]", + "answer_gt": "[0.09, 10.02]", + "answer_gt_word": "a stream of water gently flowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11827 + }, + { + "path": "YqdRkvYNjkNs.wav", + "question": "At what points does 'a stream of water gently flowing' exist completely in the timeline?", + "choice_a": "[0.37, 2.15]", + "choice_b": "[0.09, 10.02]", + "choice_c": "[6.93, 9.56]", + "choice_d": "[0.37, 2.15]", + "answer_gt": "[0.09, 10.02]", + "answer_gt_word": "a stream of water gently flowing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11828 + }, + { + "path": "YrAaEi6xKe9A.wav", + "question": "At which time markers does 'pigeons fly' fully unfold in the sound?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.67, 2.95]", + "choice_c": "[3.75, 4.81]", + "choice_d": "[7.18, 9.06]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "pigeons fly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11829 + }, + { + "path": "YraPBAKyYIGE.wav", + "question": "What are the precise second intervals of 'a small dog barks' in this clip?", + "choice_a": "[0.0, 2.53]", + "choice_b": "[7.01, 7.83]", + "choice_c": "[0.45, 0.82]", + "choice_d": "[3.57, 5.24]", + "answer_gt": "[0.45, 0.82]", + "answer_gt_word": "a small dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11830 + }, + { + "path": "Yrr5T3E4j7_Q.wav", + "question": "Where is 'a small dog barks' found in its entirety within the audio's timeline?", + "choice_a": "[0.19, 0.39]", + "choice_b": "[0.45, 0.82]", + "choice_c": "[6.13, 6.72]", + "choice_d": "[7.18, 7.4]", + "answer_gt": "[0.45, 0.82]", + "answer_gt_word": "a small dog barks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11831 + }, + { + "path": "YryTE7b5o4fE.wav", + "question": "Identify the occurrences of 'a man talks nearby' along with their time frames.", + "choice_a": "[0.38, 0.76]", + "choice_b": "[3.58, 5.73]", + "choice_c": "[7.73, 10.0]", + "choice_d": "[3.71, 8.23]", + "answer_gt": "[7.73, 10.0]", + "answer_gt_word": "a man talks nearby", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11832 + }, + { + "path": "Ys-1g0_KrLFI.wav", + "question": "How many seconds does each full presence of 'blows a piercing steam whistle' cover?", + "choice_a": "[2.65, 3.56]", + "choice_b": "[2.66, 6.61]", + "choice_c": "[3.34, 7.0]", + "choice_d": "[3.52, 8.76]", + "answer_gt": "[2.65, 3.56]", + "answer_gt_word": "blows a piercing steam whistle", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11833 + }, + { + "path": "YsM2tvbElLrI.wav", + "question": "At what timestamps does the complete 'an engine running consistently' occur?", + "choice_a": "[0.03, 10.01]", + "choice_b": "[0.03, 10.01]", + "choice_c": "[3.84, 8.7]", + "choice_d": "[7.73, 9.46]", + "answer_gt": "[0.03, 10.01]", + "answer_gt_word": "an engine running consistently", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11834 + }, + { + "path": "YsShpyu2l4YQ.wav", + "question": "At what timestamps does the complete 'an engine running consistently' occur?", + "choice_a": "[0.0, 1.96]", + "choice_b": "[0.03, 10.01]", + "choice_c": "[6.33, 7.54]", + "choice_d": "[8.85, 10.0]", + "answer_gt": "[0.03, 10.01]", + "answer_gt_word": "an engine running consistently", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11835 + }, + { + "path": "YsYj4hpDUZDQ.wav", + "question": "In what time spans is 'an engine running consistently' completely audible in this recording?", + "choice_a": "[4.18, 5.1]", + "choice_b": "[0.03, 10.01]", + "choice_c": "[0.0, 1.61]", + "choice_d": "[2.62, 3.36]", + "answer_gt": "[0.03, 10.01]", + "answer_gt_word": "an engine running consistently", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11836 + }, + { + "path": "Ysd7x_Vssqlw.wav", + "question": "How many seconds does each full presence of 'a vehicle accelerates squealing tires' cover?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[1.61, 3.18]", + "choice_c": "[2.22, 6.61]", + "choice_d": "[2.56, 7.35]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "a vehicle accelerates squealing tires", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11837 + }, + { + "path": "YsjlVMgdGSK0.wav", + "question": "Define the time intervals that fully contain 'a vehicle accelerates' in this sound.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.67, 7.77]", + "choice_d": "[4.92, 9.97]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "a vehicle accelerates", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11838 + }, + { + "path": "YsofxkNWaP0s.wav", + "question": "At what timestamps does the complete 'wind blows' occur?", + "choice_a": "[0.97, 1.62]", + "choice_b": "[4.13, 4.97]", + "choice_c": "[5.8, 6.27]", + "choice_d": "[0.0, 10.02]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11839 + }, + { + "path": "Yt8vFu8WNVU0.wav", + "question": "Define the time intervals that fully contain 'wind blows' in this sound.", + "choice_a": "[4.22, 5.98]", + "choice_b": "[9.46, 10.0]", + "choice_c": "[0.0, 10.02]", + "choice_d": "[8.1, 9.34]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "wind blows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11840 + }, + { + "path": "YtGWuKMJssrM.wav", + "question": "What second marks correspond to the entire span of 'a train horn beeps twice'?", + "choice_a": "[0.03, 0.74]", + "choice_b": "[0.02, 10.0]", + "choice_c": "[6.9, 8.89]", + "choice_d": "[6.99, 9.35]", + "answer_gt": "[0.03, 0.74]", + "answer_gt_word": "a train horn beeps twice", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11841 + }, + { + "path": "YtnUfrWBmrv8.wav", + "question": "Where is 'sirens ring' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.37, 6.85]", + "choice_c": "[1.09, 8.04]", + "choice_d": "[3.57, 8.67]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sirens ring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11842 + }, + { + "path": "YtwgRUlPxFTk.wav", + "question": "What time segments showcase 'sirens ring' in its entirety?", + "choice_a": "[1.61, 3.3]", + "choice_b": "[5.51, 7.43]", + "choice_c": "[9.26, 10.0]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sirens ring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11843 + }, + { + "path": "YufCPGoDcXLQ.wav", + "question": "What are the comprehensive timeframes of 'sirens ring' in this clip?", + "choice_a": "[0.04, 0.59]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[0.0, 0.62]", + "choice_d": "[1.27, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "sirens ring", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11844 + }, + { + "path": "YvJ7JPEFhyLA.wav", + "question": "How is 'three men talk' distributed over the sound's duration?", + "choice_a": "[3.88, 6.48]", + "choice_b": "[0.0, 9.55]", + "choice_c": "[0.0, 9.55]", + "choice_d": "[4.4, 8.81]", + "answer_gt": "[3.88, 6.48]", + "answer_gt_word": "three men talk", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11845 + }, + { + "path": "YvJ7JPEFhyLA.wav", + "question": "Where is 'men speak' consistently heard throughout the sound?", + "choice_a": "[3.9, 6.47]", + "choice_b": "[5.11, 5.58]", + "choice_c": "[6.21, 6.51]", + "choice_d": "[2.89, 5.26]", + "answer_gt": "[3.9, 6.47]", + "answer_gt_word": "men speak", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11846 + }, + { + "path": "YvNNDoUbPQHE.wav", + "question": "What are the exact times when 'a lit fuse fizzing' is present in the clip?", + "choice_a": "[0.0, 0.95]", + "choice_b": "[6.13, 10.0]", + "choice_c": "[6.47, 10.02]", + "choice_d": "[6.15, 10.0]", + "answer_gt": "[6.47, 10.02]", + "answer_gt_word": "a lit fuse fizzing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11847 + }, + { + "path": "YwI8Q_d2n68w.wav", + "question": "Where is 'a young woman speaks' found in its entirety within the audio's timeline?", + "choice_a": "[0.75, 1.48]", + "choice_b": "[3.88, 10.0]", + "choice_c": "[5.43, 10.0]", + "choice_d": "[1.31, 2.23]", + "answer_gt": "[0.75, 1.48]", + "answer_gt_word": "a young woman speaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11848 + }, + { + "path": "YwqZ135_Ssz0.wav", + "question": "How long does 'birds chirp' last each time it appears in the audio?", + "choice_a": "[0.0, 1.38]", + "choice_b": "[2.52, 4.72]", + "choice_c": "[6.42, 10.0]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "birds chirp", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11849 + }, + { + "path": "YxUIdJ_aGp-A.wav", + "question": "What are the comprehensive segments of 'an engine works nearby' in time?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[0.1, 6.62]", + "choice_c": "[1.26, 7.53]", + "choice_d": "[4.33, 7.96]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "an engine works nearby", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11850 + }, + { + "path": "YxxPndHjWTCA.wav", + "question": "What are the distinct time ranges of 'rustling' within the audio?", + "choice_a": "[0.74, 1.47]", + "choice_b": "[2.5, 3.47]", + "choice_c": "[0.0, 10.0]", + "choice_d": "[0.0, 0.15]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "rustling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11851 + }, + { + "path": "Yy0PVvrn_uGU.wav", + "question": "Enumerate the second-based occurrences of 'a train alarm' in full.", + "choice_a": "[0.01, 10.0]", + "choice_b": "[1.98, 4.27]", + "choice_c": "[6.49, 7.22]", + "choice_d": "[9.09, 10.0]", + "answer_gt": "[0.01, 10.0]", + "answer_gt_word": "a train alarm", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11852 + }, + { + "path": "Yy2ZBGpgbhHM.wav", + "question": "What are the boundaries in seconds for each 'an animal growls' presence?", + "choice_a": "[0.0, 2.38]", + "choice_b": "[1.35, 2.38]", + "choice_c": "[3.62, 4.89]", + "choice_d": "[6.53, 7.09]", + "answer_gt": "[0.0, 2.38]", + "answer_gt_word": "an animal growls", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11853 + }, + { + "path": "Yy6wsRU2aNx4.wav", + "question": "How many seconds does 'a ring bells' span in each occurrence?", + "choice_a": "[0.0, 6.19]", + "choice_b": "[0.41, 1.49]", + "choice_c": "[3.27, 3.66]", + "choice_d": "[5.92, 9.99]", + "answer_gt": "[0.0, 6.19]", + "answer_gt_word": "a ring bells", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11854 + }, + { + "path": "Yy8dSeubCN_I.wav", + "question": "How many seconds does 'a car revs and accelerates loudly' span in each occurrence?", + "choice_a": "[0.0, 2.0]", + "choice_b": "[2.3, 9.75]", + "choice_c": "[3.82, 4.14]", + "choice_d": "[7.03, 7.79]", + "answer_gt": "[0.0, 2.0]", + "answer_gt_word": "a car revs and accelerates loudly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11855 + }, + { + "path": "YyHq2AHuV2BU.wav", + "question": "Determine the full durations of 'an engine hums' in the audio file.", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 0.34]", + "choice_c": "[2.11, 3.36]", + "choice_d": "[4.22, 5.24]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "an engine hums", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11856 + }, + { + "path": "YyJ-KUG1m2CY.wav", + "question": "What time segments delineate the presence of 'a person speaks softly'?", + "choice_a": "[0.4, 1.11]", + "choice_b": "[4.74, 5.83]", + "choice_c": "[5.86, 6.66]", + "choice_d": "[2.16, 2.48]", + "answer_gt": "[0.4, 1.11]", + "answer_gt_word": "a person speaks softly", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11857 + }, + { + "path": "YyK0XUXQULZs.wav", + "question": "How are the complete segments of 'a door squeaks' timed in this audio?", + "choice_a": "[0.06, 0.79]", + "choice_b": "[9.77, 10.0]", + "choice_c": "[0.9, 2.32]", + "choice_d": "[2.82, 3.55]", + "answer_gt": "[0.06, 0.79]", + "answer_gt_word": "a door squeaks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11858 + }, + { + "path": "YyNxFTGu3TBg.wav", + "question": "What ranges in seconds capture 'a cat meows' in full?", + "choice_a": "[0.0, 8.86]", + "choice_b": "[9.75, 10.0]", + "choice_c": "[3.92, 6.44]", + "choice_d": "[1.78, 8.74]", + "answer_gt": "[3.92, 6.44]", + "answer_gt_word": "a cat meows", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11859 + }, + { + "path": "YymXRSkBSEJs.wav", + "question": "At what points does 'some brief baaing' exist completely in the timeline?", + "choice_a": "[3.98, 4.48]", + "choice_b": "[3.78, 4.46]", + "choice_c": "[4.85, 5.76]", + "choice_d": "[6.46, 10.0]", + "answer_gt": "[3.98, 4.48]", + "answer_gt_word": "some brief baaing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11860 + }, + { + "path": "YyrP1c4KzHFo.wav", + "question": "Where is 'rain falling heavily on a surface' found in its entirety within the audio's timeline?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[8.6, 9.35]", + "choice_c": "[0.95, 5.02]", + "choice_d": "[3.08, 8.39]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "rain falling heavily on a surface", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11861 + }, + { + "path": "Yysdfo6Am1jc.wav", + "question": "In what segments can 'rain falling heavily on a surface' be heard from start to finish?", + "choice_a": "[0.0, 10.02]", + "choice_b": "[5.47, 6.49]", + "choice_c": "[7.9, 9.82]", + "choice_d": "[2.03, 2.15]", + "answer_gt": "[0.0, 10.02]", + "answer_gt_word": "rain falling heavily on a surface", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11862 + }, + { + "path": "YyyGk2dAOjZM.wav", + "question": "What are the specific start and stop times of 'male speech' in the sound?", + "choice_a": "[1.84, 2.91]", + "choice_b": "[0.74, 1.57]", + "choice_c": "[3.44, 4.51]", + "choice_d": "[6.37, 7.88]", + "answer_gt": "[1.84, 2.91]", + "answer_gt_word": "male speech", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11863 + }, + { + "path": "YzB4kz0vEsTQ.wav", + "question": "What are the full time segments when 'turkeys gobbling' appears in this sound?", + "choice_a": "[0.0, 10.0]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[2.84, 3.05]", + "choice_d": "[4.91, 5.51]", + "answer_gt": "[0.0, 10.0]", + "answer_gt_word": "turkeys gobbling", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11864 + }, + { + "path": "YzV0gxh-yoqs.wav", + "question": "Where is 'a man talking' consistently heard throughout the sound?", + "choice_a": "[0.0, 4.24]", + "choice_b": "[4.8, 8.76]", + "choice_c": "[9.7, 10.0]", + "choice_d": "[8.87, 9.38]", + "answer_gt": "[8.87, 9.38]", + "answer_gt_word": "a man talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11865 + }, + { + "path": "YzguSB2RIKpw.wav", + "question": "What are the distinct time ranges of 'a girl laughing' within the audio?", + "choice_a": "[1.94, 2.18]", + "choice_b": "[2.91, 3.72]", + "choice_c": "[3.75, 4.13]", + "choice_d": "[0.0, 10.0]", + "answer_gt": "[1.94, 2.18]", + "answer_gt_word": "a girl laughing", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11866 + }, + { + "path": "YziAl_IGwMpM.wav", + "question": "Enumerate the second-based occurrences of 'a helicopter engine starting up' in full.", + "choice_a": "[0.0, 4.96]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[1.61, 4.71]", + "choice_d": "[2.34, 5.72]", + "answer_gt": "[0.0, 4.96]", + "answer_gt_word": "a helicopter engine starting up", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11867 + }, + { + "path": "YziQzaDcMzJg.wav", + "question": "What are the specific start and stop times of 'a woman talking' in the sound?", + "choice_a": "[0.0, 0.34]", + "choice_b": "[1.07, 2.1]", + "choice_c": "[6.86, 7.53]", + "choice_d": "[5.24, 6.87]", + "answer_gt": "[6.86, 7.53]", + "answer_gt_word": "a woman talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11868 + }, + { + "path": "YzlJ8jLkCRUU.wav", + "question": "What second marks correspond to the entire span of 'a woman talking'?", + "choice_a": "[0.0, 0.84]", + "choice_b": "[5.52, 6.18]", + "choice_c": "[6.66, 7.41]", + "choice_d": "[6.86, 7.53]", + "answer_gt": "[6.86, 7.53]", + "answer_gt_word": "a woman talking", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11869 + }, + { + "path": "YzlL_aOhD_X0.wav", + "question": "Which sections of the sound contain 'dogs are whimpering' in their entirety?", + "choice_a": "[0.03, 9.99]", + "choice_b": "[0.0, 10.0]", + "choice_c": "[7.08, 9.51]", + "choice_d": "[8.91, 9.99]", + "answer_gt": "[0.03, 9.99]", + "answer_gt_word": "dogs are whimpering", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11870 + }, + { + "path": "YzpsOHpUwUfY.wav", + "question": "What are the full time segments when 'a train horn honks' appears in this sound?", + "choice_a": "[3.29, 3.84]", + "choice_b": "[2.47, 3.91]", + "choice_c": "[2.67, 4.16]", + "choice_d": "[3.86, 8.73]", + "answer_gt": "[3.29, 3.84]", + "answer_gt_word": "a train horn honks", + "task_name": "Audio_Grounding", + "dataset_name": "AudioGrounding", + "uniq_id": 11871 + }, + { + "path": "f0638_0_sneeze.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11872 + }, + { + "path": "f0592_0_sigh.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11873 + }, + { + "path": "m3072_0_sniff.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11874 + }, + { + "path": "f0383_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11875 + }, + { + "path": "m0283_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11876 + }, + { + "path": "m1462_0_sniff.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11877 + }, + { + "path": "m2415_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11878 + }, + { + "path": "f3019_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11879 + }, + { + "path": "m1642_0_sigh.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11880 + }, + { + "path": "f2385_0_throatclearing.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11881 + }, + { + "path": "f1941_1_sigh.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11882 + }, + { + "path": "m1403_0_sniff.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11883 + }, + { + "path": "f1086_0_sneeze.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11884 + }, + { + "path": "m3004_0_sniff.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11885 + }, + { + "path": "f2183_0_sniff.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11886 + }, + { + "path": "f1646_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11887 + }, + { + "path": "f3184_0_cough.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11888 + }, + { + "path": "m1735_0_sniff.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11889 + }, + { + "path": "m0247_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11890 + }, + { + "path": "f1601_0_throatclearing.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11891 + }, + { + "path": "f0510_0_sniff.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11892 + }, + { + "path": "f0991_0_sigh.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11893 + }, + { + "path": "m1049_0_laughter.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11894 + }, + { + "path": "o1900_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11895 + }, + { + "path": "f3116_0_laughter.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11896 + }, + { + "path": "f0116_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11897 + }, + { + "path": "f1392_0_sneeze.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11898 + }, + { + "path": "f2890_0_sigh.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11899 + }, + { + "path": "f2749_0_sniff.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11900 + }, + { + "path": "m0283_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11901 + }, + { + "path": "m3176_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11902 + }, + { + "path": "f2462_0_laughter.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11903 + }, + { + "path": "m1187_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11904 + }, + { + "path": "m2803_0_sneeze.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11905 + }, + { + "path": "f1964_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11906 + }, + { + "path": "f1153_0_laughter.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11907 + }, + { + "path": "f2984_0_throatclearing.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11908 + }, + { + "path": "f1435_0_sigh.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11909 + }, + { + "path": "m0895_0_sniff.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11910 + }, + { + "path": "f0422_0_laughter.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11911 + }, + { + "path": "m2270_0_cough.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11912 + }, + { + "path": "m2539_0_sigh.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11913 + }, + { + "path": "m2713_0_sigh.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11914 + }, + { + "path": "m2504_0_sniff.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11915 + }, + { + "path": "f0124_0_sneeze.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11916 + }, + { + "path": "f2666_0_laughter.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11917 + }, + { + "path": "f1957_2_sigh.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11918 + }, + { + "path": "f0930_0_sigh.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11919 + }, + { + "path": "f0124_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11920 + }, + { + "path": "m0526_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11921 + }, + { + "path": "m1330_0_sigh.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11922 + }, + { + "path": "m1609_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11923 + }, + { + "path": "f2207_0_laughter.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11924 + }, + { + "path": "m2759_0_laughter.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11925 + }, + { + "path": "f2122_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11926 + }, + { + "path": "f0401_0_sigh.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11927 + }, + { + "path": "m1105_0_sneeze.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11928 + }, + { + "path": "m0557_0_laughter.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11929 + }, + { + "path": "m2418_0_cough.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11930 + }, + { + "path": "f1414_0_sneeze.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11931 + }, + { + "path": "f0194_0_cough.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11932 + }, + { + "path": "f1585_0_sneeze.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11933 + }, + { + "path": "m0778_0_cough.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11934 + }, + { + "path": "m2150_0_sneeze.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11935 + }, + { + "path": "m2270_0_throatclearing.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11936 + }, + { + "path": "f2912_0_laughter.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11937 + }, + { + "path": "f0463_0_sneeze.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11938 + }, + { + "path": "f0556_0_sigh.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11939 + }, + { + "path": "f0208_0_sigh.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11940 + }, + { + "path": "f0661_0_laughter.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11941 + }, + { + "path": "m1187_0_laughter.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11942 + }, + { + "path": "f2354_0_throatclearing.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11943 + }, + { + "path": "f1252_0_sigh.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11944 + }, + { + "path": "m1049_0_cough.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11945 + }, + { + "path": "f1322_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11946 + }, + { + "path": "f1290_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11947 + }, + { + "path": "f1405_0_cough.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11948 + }, + { + "path": "m1168_0_cough.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11949 + }, + { + "path": "m2499_0_sniff.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11950 + }, + { + "path": "m0420_0_laughter.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11951 + }, + { + "path": "f1971_0_sniff.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11952 + }, + { + "path": "f1023_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11953 + }, + { + "path": "m0431_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11954 + }, + { + "path": "m1567_0_cough.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11955 + }, + { + "path": "m2372_0_throatclearing.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11956 + }, + { + "path": "f1298_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11957 + }, + { + "path": "m2277_0_throatclearing.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11958 + }, + { + "path": "f2787_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11959 + }, + { + "path": "f1941_1_throatclearing.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11960 + }, + { + "path": "f1741_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11961 + }, + { + "path": "m2081_0_laughter.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11962 + }, + { + "path": "f2739_1_sneeze.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11963 + }, + { + "path": "m3292_0_throatclearing.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11964 + }, + { + "path": "o2139_0_sneeze.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11965 + }, + { + "path": "m0210_0_sigh.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11966 + }, + { + "path": "m1933_0_sniff.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11967 + }, + { + "path": "f1401_0_cough.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11968 + }, + { + "path": "f1741_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11969 + }, + { + "path": "f3249_0_sneeze.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11970 + }, + { + "path": "f1145_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11971 + }, + { + "path": "m0561_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11972 + }, + { + "path": "m2081_0_sigh.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11973 + }, + { + "path": "m0768_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11974 + }, + { + "path": "f1813_0_sniff.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11975 + }, + { + "path": "f0969_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11976 + }, + { + "path": "m2721_0_sneeze.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11977 + }, + { + "path": "m1134_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11978 + }, + { + "path": "f1839_0_cough.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11979 + }, + { + "path": "m0787_0_sigh.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11980 + }, + { + "path": "m2339_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11981 + }, + { + "path": "m2702_0_sniff.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11982 + }, + { + "path": "f0991_0_cough.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11983 + }, + { + "path": "m1783_0_throatclearing.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11984 + }, + { + "path": "m2468_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11985 + }, + { + "path": "f1008_0_laughter.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11986 + }, + { + "path": "f1964_1_sniff.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11987 + }, + { + "path": "m0670_0_laughter.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11988 + }, + { + "path": "f2643_0_cough.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11989 + }, + { + "path": "m2277_0_sigh.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11990 + }, + { + "path": "f1298_0_laughter.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11991 + }, + { + "path": "f0810_0_cough.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11992 + }, + { + "path": "f1585_0_sniff.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11993 + }, + { + "path": "f3257_0_throatclearing.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11994 + }, + { + "path": "f0791_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11995 + }, + { + "path": "f2248_0_laughter.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11996 + }, + { + "path": "f1366_0_sneeze.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11997 + }, + { + "path": "m0434_0_cough.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11998 + }, + { + "path": "f3174_0_throatclearing.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 11999 + }, + { + "path": "f1086_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12000 + }, + { + "path": "m2999_0_sigh.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12001 + }, + { + "path": "m2304_0_sneeze.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12002 + }, + { + "path": "m0026_0_sneeze.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12003 + }, + { + "path": "m0127_0_sneeze.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12004 + }, + { + "path": "m2004_0_sneeze.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12005 + }, + { + "path": "f1813_0_sigh.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12006 + }, + { + "path": "m1563_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12007 + }, + { + "path": "f0969_0_laughter.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12008 + }, + { + "path": "m2490_0_sigh.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12009 + }, + { + "path": "m1783_0_cough.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12010 + }, + { + "path": "m1444_0_laughter.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12011 + }, + { + "path": "f0788_0_sneeze.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12012 + }, + { + "path": "m0817_0_sigh.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12013 + }, + { + "path": "f2354_0_sneeze.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12014 + }, + { + "path": "f2118_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12015 + }, + { + "path": "f0605_0_sneeze.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12016 + }, + { + "path": "m2493_0_cough.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12017 + }, + { + "path": "m1756_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12018 + }, + { + "path": "f3084_0_laughter.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12019 + }, + { + "path": "f1438_0_throatclearing.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12020 + }, + { + "path": "m1579_0_throatclearing.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12021 + }, + { + "path": "m0855_0_sniff.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12022 + }, + { + "path": "m1543_0_laughter.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12023 + }, + { + "path": "m1891_0_laughter.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12024 + }, + { + "path": "f0631_0_sneeze.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12025 + }, + { + "path": "f1941_0_throatclearing.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12026 + }, + { + "path": "f0167_0_sneeze.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12027 + }, + { + "path": "f0606_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12028 + }, + { + "path": "f1721_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12029 + }, + { + "path": "f2192_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12030 + }, + { + "path": "f1396_0_laughter.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12031 + }, + { + "path": "f2092_0_sneeze.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12032 + }, + { + "path": "o1897_0_sneeze.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12033 + }, + { + "path": "m1399_0_laughter.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12034 + }, + { + "path": "m2881_0_sniff.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12035 + }, + { + "path": "f1321_0_sigh.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12036 + }, + { + "path": "f0091_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12037 + }, + { + "path": "f0556_0_sniff.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12038 + }, + { + "path": "f1562_0_sneeze.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12039 + }, + { + "path": "m0075_0_cough.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12040 + }, + { + "path": "f0755_0_laughter.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12041 + }, + { + "path": "m2753_0_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12042 + }, + { + "path": "m2493_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12043 + }, + { + "path": "f0027_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12044 + }, + { + "path": "m1543_0_sneeze.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12045 + }, + { + "path": "f2443_0_sigh.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12046 + }, + { + "path": "f0087_0_cough.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12047 + }, + { + "path": "m0903_0_cough.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12048 + }, + { + "path": "m0768_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12049 + }, + { + "path": "m0670_0_cough.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12050 + }, + { + "path": "m1806_0_laughter.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12051 + }, + { + "path": "m1195_0_throatclearing.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12052 + }, + { + "path": "m3238_0_laughter.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12053 + }, + { + "path": "f2336_0_sniff.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12054 + }, + { + "path": "m1347_0_laughter.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12055 + }, + { + "path": "m2238_0_sneeze.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12056 + }, + { + "path": "m0022_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12057 + }, + { + "path": "f2293_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12058 + }, + { + "path": "m1180_0_sneeze.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12059 + }, + { + "path": "f1547_0_throatclearing.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12060 + }, + { + "path": "m0754_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12061 + }, + { + "path": "m1403_0_sigh.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12062 + }, + { + "path": "m0289_0_sniff.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12063 + }, + { + "path": "m2222_0_sniff.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12064 + }, + { + "path": "m2258_0_laughter.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12065 + }, + { + "path": "m1708_0_sigh.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12066 + }, + { + "path": "m1735_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12067 + }, + { + "path": "f2207_0_throatclearing.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12068 + }, + { + "path": "f1473_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12069 + }, + { + "path": "m1251_0_sniff.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12070 + }, + { + "path": "f2937_0_sniff.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12071 + }, + { + "path": "f2764_0_laughter.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12072 + }, + { + "path": "m1770_0_laughter.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12073 + }, + { + "path": "m1892_0_sniff.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12074 + }, + { + "path": "m2545_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12075 + }, + { + "path": "f2164_0_sneeze.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12076 + }, + { + "path": "f0258_0_sniff.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12077 + }, + { + "path": "m1553_0_sniff.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12078 + }, + { + "path": "f1941_0_sigh.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12079 + }, + { + "path": "m1207_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12080 + }, + { + "path": "m2339_0_cough.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12081 + }, + { + "path": "f1590_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12082 + }, + { + "path": "f1719_0_sneeze.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12083 + }, + { + "path": "f2890_0_cough.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12084 + }, + { + "path": "m2559_0_sneeze.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12085 + }, + { + "path": "f0116_0_laughter.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12086 + }, + { + "path": "m0308_0_sigh.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12087 + }, + { + "path": "f0607_2_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12088 + }, + { + "path": "o1900_0_throatclearing.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12089 + }, + { + "path": "m2999_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12090 + }, + { + "path": "f2459_0_throatclearing.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12091 + }, + { + "path": "f1706_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12092 + }, + { + "path": "m0871_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12093 + }, + { + "path": "f2438_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12094 + }, + { + "path": "m1203_0_throatclearing.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12095 + }, + { + "path": "m3250_0_laughter.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12096 + }, + { + "path": "f2694_0_cough.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12097 + }, + { + "path": "f1691_0_cough.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12098 + }, + { + "path": "m0561_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12099 + }, + { + "path": "m1444_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12100 + }, + { + "path": "m1409_0_sneeze.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12101 + }, + { + "path": "m0825_0_cough.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12102 + }, + { + "path": "f2462_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12103 + }, + { + "path": "f2275_0_throatclearing.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12104 + }, + { + "path": "m0376_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12105 + }, + { + "path": "f2257_0_sigh.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12106 + }, + { + "path": "f1113_0_sneeze.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12107 + }, + { + "path": "m0390_0_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12108 + }, + { + "path": "f0969_0_sneeze.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12109 + }, + { + "path": "f3352_0_throatclearing.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12110 + }, + { + "path": "f0004_0_sneeze.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12111 + }, + { + "path": "o1897_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12112 + }, + { + "path": "m0831_0_laughter.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12113 + }, + { + "path": "f3019_0_cough.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12114 + }, + { + "path": "m2025_0_sniff.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12115 + }, + { + "path": "f1467_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12116 + }, + { + "path": "m1012_0_sniff.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12117 + }, + { + "path": "f1987_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12118 + }, + { + "path": "f0418_0_laughter.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12119 + }, + { + "path": "m1126_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12120 + }, + { + "path": "m2024_0_sneeze.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12121 + }, + { + "path": "f2750_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12122 + }, + { + "path": "f1321_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12123 + }, + { + "path": "f0968_0_sniff.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12124 + }, + { + "path": "m2600_0_cough.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12125 + }, + { + "path": "f2938_0_sniff.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12126 + }, + { + "path": "m0344_0_sigh.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12127 + }, + { + "path": "m2634_0_sniff.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12128 + }, + { + "path": "m2759_0_sigh.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12129 + }, + { + "path": "m1910_0_sniff.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12130 + }, + { + "path": "m0308_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12131 + }, + { + "path": "m2970_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12132 + }, + { + "path": "m2283_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12133 + }, + { + "path": "m1093_0_sniff.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12134 + }, + { + "path": "f0623_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12135 + }, + { + "path": "f0332_0_laughter.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12136 + }, + { + "path": "f2686_0_laughter.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12137 + }, + { + "path": "m1489_0_cough.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12138 + }, + { + "path": "f0292_0_sigh.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12139 + }, + { + "path": "m2611_0_sniff.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12140 + }, + { + "path": "m2221_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12141 + }, + { + "path": "f2186_0_laughter.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12142 + }, + { + "path": "m3025_0_sniff.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12143 + }, + { + "path": "m2389_0_sneeze.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12144 + }, + { + "path": "f0933_0_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12145 + }, + { + "path": "m1892_0_sigh.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12146 + }, + { + "path": "m0946_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12147 + }, + { + "path": "f2470_0_laughter.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12148 + }, + { + "path": "m1338_0_laughter.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12149 + }, + { + "path": "m1208_0_laughter.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12150 + }, + { + "path": "m1029_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12151 + }, + { + "path": "m1131_0_sigh.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12152 + }, + { + "path": "m2559_0_sigh.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12153 + }, + { + "path": "m1797_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12154 + }, + { + "path": "m1439_0_cough.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12155 + }, + { + "path": "f0613_0_throatclearing.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12156 + }, + { + "path": "m1203_0_laughter.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12157 + }, + { + "path": "m2930_0_sigh.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12158 + }, + { + "path": "f1321_0_throatclearing.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12159 + }, + { + "path": "f2155_0_throatclearing.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12160 + }, + { + "path": "f1921_0_sniff.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12161 + }, + { + "path": "f0615_0_laughter.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12162 + }, + { + "path": "m2954_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12163 + }, + { + "path": "f1349_1_laughter.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12164 + }, + { + "path": "m3013_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12165 + }, + { + "path": "f2847_0_sneeze.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12166 + }, + { + "path": "m1164_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12167 + }, + { + "path": "m1925_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12168 + }, + { + "path": "m3176_0_sigh.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12169 + }, + { + "path": "m1528_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12170 + }, + { + "path": "f0516_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12171 + }, + { + "path": "m1869_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12172 + }, + { + "path": "f1974_0_sneeze.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12173 + }, + { + "path": "m1869_0_sigh.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12174 + }, + { + "path": "f2155_0_sigh.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12175 + }, + { + "path": "m0730_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12176 + }, + { + "path": "f2365_0_sigh.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12177 + }, + { + "path": "f1796_0_laughter.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12178 + }, + { + "path": "f0651_0_sigh.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12179 + }, + { + "path": "f1438_0_cough.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12180 + }, + { + "path": "m2596_0_throatclearing.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12181 + }, + { + "path": "f2746_0_laughter.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12182 + }, + { + "path": "m1708_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12183 + }, + { + "path": "m1093_0_throatclearing.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12184 + }, + { + "path": "m2565_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12185 + }, + { + "path": "m1168_0_throatclearing.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12186 + }, + { + "path": "m0733_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12187 + }, + { + "path": "f1113_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12188 + }, + { + "path": "m0846_0_throatclearing.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12189 + }, + { + "path": "m0374_0_laughter.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12190 + }, + { + "path": "f0810_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12191 + }, + { + "path": "m2751_0_sniff.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12192 + }, + { + "path": "m0095_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12193 + }, + { + "path": "m1484_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12194 + }, + { + "path": "f0087_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12195 + }, + { + "path": "m0470_0_sniff.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12196 + }, + { + "path": "m2353_0_sneeze.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12197 + }, + { + "path": "f1332_0_sneeze.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12198 + }, + { + "path": "m1845_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12199 + }, + { + "path": "m3120_0_throatclearing.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12200 + }, + { + "path": "m1756_1_cough.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12201 + }, + { + "path": "f0276_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12202 + }, + { + "path": "m1208_0_throatclearing.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12203 + }, + { + "path": "m0836_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12204 + }, + { + "path": "m0202_0_cough.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12205 + }, + { + "path": "f1931_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12206 + }, + { + "path": "o1900_0_sigh.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12207 + }, + { + "path": "f1839_0_sniff.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12208 + }, + { + "path": "m0526_0_sniff.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12209 + }, + { + "path": "m0127_0_sniff.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12210 + }, + { + "path": "m3253_0_laughter.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12211 + }, + { + "path": "m0198_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12212 + }, + { + "path": "f0791_0_sigh.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12213 + }, + { + "path": "f2170_0_throatclearing.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12214 + }, + { + "path": "f2784_0_laughter.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12215 + }, + { + "path": "f1430_0_throatclearing.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12216 + }, + { + "path": "f1971_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12217 + }, + { + "path": "f2183_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12218 + }, + { + "path": "f0226_0_cough.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12219 + }, + { + "path": "m0596_0_laughter.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12220 + }, + { + "path": "f2003_0_sigh.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12221 + }, + { + "path": "f3149_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12222 + }, + { + "path": "m2490_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12223 + }, + { + "path": "m2559_0_sniff.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12224 + }, + { + "path": "f2170_0_sneeze.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12225 + }, + { + "path": "m1609_1_sniff.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12226 + }, + { + "path": "f1746_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12227 + }, + { + "path": "f0029_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12228 + }, + { + "path": "m2140_0_laughter.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12229 + }, + { + "path": "m2258_0_sigh.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12230 + }, + { + "path": "f0942_0_sigh.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12231 + }, + { + "path": "f3126_0_sniff.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12232 + }, + { + "path": "m2081_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12233 + }, + { + "path": "m3289_0_laughter.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12234 + }, + { + "path": "f3309_0_sniff.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12235 + }, + { + "path": "f3174_0_laughter.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12236 + }, + { + "path": "f1590_0_laughter.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12237 + }, + { + "path": "m3289_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12238 + }, + { + "path": "m2651_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12239 + }, + { + "path": "m1187_0_cough.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12240 + }, + { + "path": "f1156_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12241 + }, + { + "path": "f0781_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12242 + }, + { + "path": "m1869_0_laughter.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12243 + }, + { + "path": "m2082_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12244 + }, + { + "path": "m2177_0_cough.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12245 + }, + { + "path": "m2509_0_laughter.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12246 + }, + { + "path": "m1287_0_laughter.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12247 + }, + { + "path": "f1778_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12248 + }, + { + "path": "f2503_0_sneeze.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12249 + }, + { + "path": "m1114_0_sneeze.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12250 + }, + { + "path": "f3126_0_laughter.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12251 + }, + { + "path": "f0933_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12252 + }, + { + "path": "f1405_0_throatclearing.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12253 + }, + { + "path": "f0226_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12254 + }, + { + "path": "m0198_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12255 + }, + { + "path": "f1668_0_sneeze.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12256 + }, + { + "path": "m0325_0_sneeze.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12257 + }, + { + "path": "m1579_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12258 + }, + { + "path": "f0835_0_laughter.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12259 + }, + { + "path": "f3334_0_laughter.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12260 + }, + { + "path": "m1672_0_laughter.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12261 + }, + { + "path": "m0730_0_sniff.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12262 + }, + { + "path": "m0861_0_laughter.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12263 + }, + { + "path": "f0182_0_laughter.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12264 + }, + { + "path": "f1349_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12265 + }, + { + "path": "m1906_0_laughter.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12266 + }, + { + "path": "m2072_0_cough.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12267 + }, + { + "path": "f2186_0_cough.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12268 + }, + { + "path": "f0181_0_sigh.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12269 + }, + { + "path": "f3106_0_throatclearing.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12270 + }, + { + "path": "m3101_0_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12271 + }, + { + "path": "f0866_0_sniff.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12272 + }, + { + "path": "f2799_0_sigh.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12273 + }, + { + "path": "f1719_0_sniff.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12274 + }, + { + "path": "f1349_0_sigh.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12275 + }, + { + "path": "m1583_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12276 + }, + { + "path": "m0047_0_sniff.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12277 + }, + { + "path": "m2384_0_sneeze.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12278 + }, + { + "path": "f2443_0_throatclearing.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12279 + }, + { + "path": "f3264_0_sigh.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12280 + }, + { + "path": "f1796_0_sniff.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12281 + }, + { + "path": "m2297_1_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12282 + }, + { + "path": "m3101_0_cough.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12283 + }, + { + "path": "f3115_0_sneeze.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12284 + }, + { + "path": "f0151_0_sniff.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12285 + }, + { + "path": "f0593_2_sigh.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12286 + }, + { + "path": "m1417_0_sigh.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12287 + }, + { + "path": "m1021_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12288 + }, + { + "path": "m0538_0_sneeze.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12289 + }, + { + "path": "m3242_0_laughter.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12290 + }, + { + "path": "f3324_0_laughter.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12291 + }, + { + "path": "f1798_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12292 + }, + { + "path": "m1955_0_laughter.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12293 + }, + { + "path": "m2389_0_cough.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12294 + }, + { + "path": "f0496_0_sigh.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12295 + }, + { + "path": "m2262_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12296 + }, + { + "path": "f2837_0_sigh.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12297 + }, + { + "path": "f1305_0_sniff.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12298 + }, + { + "path": "m2754_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12299 + }, + { + "path": "f3215_0_laughter.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12300 + }, + { + "path": "m1093_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12301 + }, + { + "path": "f0743_0_sniff.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12302 + }, + { + "path": "m2304_0_sniff.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12303 + }, + { + "path": "f0194_0_laughter.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12304 + }, + { + "path": "f1942_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12305 + }, + { + "path": "m2432_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12306 + }, + { + "path": "f1884_0_laughter.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12307 + }, + { + "path": "f2187_0_sneeze.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12308 + }, + { + "path": "f1334_0_cough.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12309 + }, + { + "path": "f1132_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12310 + }, + { + "path": "m1005_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12311 + }, + { + "path": "f0298_0_throatclearing.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12312 + }, + { + "path": "m0595_1_sneeze.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12313 + }, + { + "path": "m2968_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12314 + }, + { + "path": "m2339_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12315 + }, + { + "path": "m2998_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12316 + }, + { + "path": "m1972_0_sneeze.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12317 + }, + { + "path": "m1673_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12318 + }, + { + "path": "m2768_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12319 + }, + { + "path": "m0886_0_sigh.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12320 + }, + { + "path": "m0064_0_laughter.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12321 + }, + { + "path": "f1220_0_cough.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12322 + }, + { + "path": "f0889_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12323 + }, + { + "path": "m3074_0_sigh.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12324 + }, + { + "path": "f1315_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12325 + }, + { + "path": "f2861_0_sigh.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12326 + }, + { + "path": "f0613_0_sneeze.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12327 + }, + { + "path": "f3063_0_throatclearing.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12328 + }, + { + "path": "m0596_0_cough.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12329 + }, + { + "path": "m1911_0_cough.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12330 + }, + { + "path": "f0029_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12331 + }, + { + "path": "m2274_0_cough.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12332 + }, + { + "path": "m3292_0_laughter.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12333 + }, + { + "path": "f1305_0_laughter.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12334 + }, + { + "path": "f2847_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12335 + }, + { + "path": "m2733_1_sneeze.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12336 + }, + { + "path": "m2339_0_sigh.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12337 + }, + { + "path": "f0651_0_cough.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12338 + }, + { + "path": "f1346_0_sneeze.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12339 + }, + { + "path": "m3212_0_throatclearing.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12340 + }, + { + "path": "f0604_0_laughter.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12341 + }, + { + "path": "m2639_0_sigh.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12342 + }, + { + "path": "m2200_0_cough.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12343 + }, + { + "path": "m1442_0_cough.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12344 + }, + { + "path": "f1435_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12345 + }, + { + "path": "f0237_0_laughter.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12346 + }, + { + "path": "f1254_0_sigh.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12347 + }, + { + "path": "f0556_0_sneeze.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12348 + }, + { + "path": "f1470_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12349 + }, + { + "path": "f1242_1_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12350 + }, + { + "path": "f2796_0_cough.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12351 + }, + { + "path": "f0509_0_laughter.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12352 + }, + { + "path": "f2293_0_sigh.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12353 + }, + { + "path": "m1732_0_sneeze.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12354 + }, + { + "path": "m1557_0_sigh.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12355 + }, + { + "path": "m1537_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12356 + }, + { + "path": "m1979_0_laughter.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12357 + }, + { + "path": "m3197_0_sniff.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12358 + }, + { + "path": "f0182_0_sigh.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12359 + }, + { + "path": "f2170_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12360 + }, + { + "path": "m3025_0_sneeze.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12361 + }, + { + "path": "f0200_0_sniff.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12362 + }, + { + "path": "m1436_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12363 + }, + { + "path": "m0731_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12364 + }, + { + "path": "m2392_0_cough.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12365 + }, + { + "path": "m1178_0_laughter.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12366 + }, + { + "path": "f1305_0_cough.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12367 + }, + { + "path": "m2547_0_sigh.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12368 + }, + { + "path": "f1054_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12369 + }, + { + "path": "m0561_1_sniff.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12370 + }, + { + "path": "f0509_0_sneeze.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12371 + }, + { + "path": "f0151_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12372 + }, + { + "path": "f1971_0_throatclearing.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12373 + }, + { + "path": "f2814_0_laughter.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12374 + }, + { + "path": "f0755_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12375 + }, + { + "path": "f3352_0_laughter.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12376 + }, + { + "path": "m1532_0_cough.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12377 + }, + { + "path": "m2277_0_laughter.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12378 + }, + { + "path": "m2710_0_cough.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12379 + }, + { + "path": "m1583_0_sniff.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12380 + }, + { + "path": "m0431_0_laughter.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12381 + }, + { + "path": "f0285_0_throatclearing.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12382 + }, + { + "path": "f2725_0_cough.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12383 + }, + { + "path": "m0672_0_sigh.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12384 + }, + { + "path": "m3320_0_throatclearing.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12385 + }, + { + "path": "f0405_0_cough.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12386 + }, + { + "path": "m3203_0_sigh.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12387 + }, + { + "path": "f1273_0_laughter.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12388 + }, + { + "path": "m3120_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12389 + }, + { + "path": "f1686_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12390 + }, + { + "path": "f0909_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12391 + }, + { + "path": "m1543_0_sigh.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12392 + }, + { + "path": "m1557_0_sniff.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12393 + }, + { + "path": "f1317_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12394 + }, + { + "path": "f0968_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12395 + }, + { + "path": "f2003_0_sniff.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12396 + }, + { + "path": "f0860_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12397 + }, + { + "path": "f2187_0_laughter.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12398 + }, + { + "path": "f2966_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12399 + }, + { + "path": "m1135_0_cough.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12400 + }, + { + "path": "f2257_0_sniff.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12401 + }, + { + "path": "f0461_0_laughter.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12402 + }, + { + "path": "f1620_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12403 + }, + { + "path": "f1334_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12404 + }, + { + "path": "f2984_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12405 + }, + { + "path": "f0605_1_sniff.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12406 + }, + { + "path": "f0788_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12407 + }, + { + "path": "m1938_0_cough.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12408 + }, + { + "path": "m2512_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12409 + }, + { + "path": "f0605_1_sneeze.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12410 + }, + { + "path": "m0092_0_laughter.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12411 + }, + { + "path": "m2432_0_cough.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12412 + }, + { + "path": "m2710_0_laughter.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12413 + }, + { + "path": "f0661_0_cough.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12414 + }, + { + "path": "m2702_0_laughter.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12415 + }, + { + "path": "m1916_0_cough.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12416 + }, + { + "path": "m0921_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12417 + }, + { + "path": "o0338_0_sneeze.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12418 + }, + { + "path": "m2283_0_laughter.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12419 + }, + { + "path": "f1007_0_throatclearing.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12420 + }, + { + "path": "m2768_1_sigh.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12421 + }, + { + "path": "m2767_0_sniff.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12422 + }, + { + "path": "m3014_0_cough.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12423 + }, + { + "path": "f0860_0_laughter.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12424 + }, + { + "path": "f0607_3_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12425 + }, + { + "path": "f1829_0_throatclearing.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12426 + }, + { + "path": "f2183_0_throatclearing.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12427 + }, + { + "path": "m2406_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12428 + }, + { + "path": "m2372_0_sneeze.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12429 + }, + { + "path": "m0805_0_sniff.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12430 + }, + { + "path": "m3244_0_sniff.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12431 + }, + { + "path": "f0607_0_sniff.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12432 + }, + { + "path": "m0861_0_sigh.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12433 + }, + { + "path": "f2890_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12434 + }, + { + "path": "m1275_0_sniff.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12435 + }, + { + "path": "f1108_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12436 + }, + { + "path": "m1673_0_laughter.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12437 + }, + { + "path": "m1891_0_throatclearing.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12438 + }, + { + "path": "f1585_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12439 + }, + { + "path": "m1542_0_sigh.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12440 + }, + { + "path": "f0509_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12441 + }, + { + "path": "m2919_0_sigh.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12442 + }, + { + "path": "m2353_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12443 + }, + { + "path": "f3215_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12444 + }, + { + "path": "m2069_1_laughter.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12445 + }, + { + "path": "f1073_1_laughter.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12446 + }, + { + "path": "f1941_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12447 + }, + { + "path": "m2788_0_throatclearing.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12448 + }, + { + "path": "m0733_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12449 + }, + { + "path": "f0710_0_sigh.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12450 + }, + { + "path": "m1546_0_sniff.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12451 + }, + { + "path": "f0276_0_cough.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12452 + }, + { + "path": "f1685_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12453 + }, + { + "path": "m3292_0_sniff.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12454 + }, + { + "path": "m1845_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12455 + }, + { + "path": "f1796_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12456 + }, + { + "path": "f2023_0_sniff.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12457 + }, + { + "path": "f2959_0_laughter.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12458 + }, + { + "path": "f2388_0_sigh.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12459 + }, + { + "path": "m0595_2_throatclearing.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12460 + }, + { + "path": "m2174_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12461 + }, + { + "path": "f0969_0_cough.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12462 + }, + { + "path": "m1537_0_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12463 + }, + { + "path": "m2304_0_laughter.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12464 + }, + { + "path": "m3250_0_throatclearing.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12465 + }, + { + "path": "f2293_1_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12466 + }, + { + "path": "f2643_0_laughter.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12467 + }, + { + "path": "m2340_0_cough.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12468 + }, + { + "path": "f1434_0_sniff.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12469 + }, + { + "path": "m2335_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12470 + }, + { + "path": "f0889_0_cough.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12471 + }, + { + "path": "f2438_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12472 + }, + { + "path": "f1480_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12473 + }, + { + "path": "m0137_0_sniff.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12474 + }, + { + "path": "m2545_0_sniff.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12475 + }, + { + "path": "f0096_0_sniff.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12476 + }, + { + "path": "m0801_0_sigh.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12477 + }, + { + "path": "m1471_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12478 + }, + { + "path": "m3013_0_cough.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12479 + }, + { + "path": "m1869_0_sneeze.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12480 + }, + { + "path": "f1829_0_sneeze.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12481 + }, + { + "path": "m3072_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12482 + }, + { + "path": "m0808_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12483 + }, + { + "path": "f1073_0_sigh.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12484 + }, + { + "path": "m2896_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12485 + }, + { + "path": "m0946_0_cough.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12486 + }, + { + "path": "m1824_0_cough.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12487 + }, + { + "path": "f1273_0_sigh.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12488 + }, + { + "path": "m2281_0_sigh.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12489 + }, + { + "path": "m1250_0_sneeze.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12490 + }, + { + "path": "m2708_0_sniff.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12491 + }, + { + "path": "m1823_0_sneeze.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12492 + }, + { + "path": "f1961_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12493 + }, + { + "path": "m1067_0_laughter.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12494 + }, + { + "path": "m0771_0_laughter.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12495 + }, + { + "path": "f0281_0_sigh.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12496 + }, + { + "path": "m2648_0_laughter.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12497 + }, + { + "path": "f1691_0_sigh.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12498 + }, + { + "path": "m2509_0_cough.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12499 + }, + { + "path": "m2069_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12500 + }, + { + "path": "m0420_0_cough.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12501 + }, + { + "path": "f0852_0_laughter.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12502 + }, + { + "path": "f3053_0_sneeze.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12503 + }, + { + "path": "m0831_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12504 + }, + { + "path": "m2710_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12505 + }, + { + "path": "f1242_0_sneeze.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12506 + }, + { + "path": "f3055_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12507 + }, + { + "path": "m2509_0_throatclearing.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12508 + }, + { + "path": "f1957_1_laughter.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12509 + }, + { + "path": "f0858_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12510 + }, + { + "path": "m1824_0_sniff.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12511 + }, + { + "path": "f0604_0_sniff.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12512 + }, + { + "path": "f2608_0_cough.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12513 + }, + { + "path": "f1654_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12514 + }, + { + "path": "m0366_0_throatclearing.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12515 + }, + { + "path": "f1957_2_sniff.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12516 + }, + { + "path": "f0237_0_sneeze.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12517 + }, + { + "path": "m3320_0_sigh.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12518 + }, + { + "path": "f0401_0_throatclearing.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12519 + }, + { + "path": "f2537_0_cough.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12520 + }, + { + "path": "m0164_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12521 + }, + { + "path": "f2441_0_sneeze.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12522 + }, + { + "path": "f2608_0_sniff.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12523 + }, + { + "path": "m1173_0_sigh.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12524 + }, + { + "path": "f3215_0_throatclearing.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12525 + }, + { + "path": "m0932_0_sniff.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12526 + }, + { + "path": "f1811_0_sniff.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12527 + }, + { + "path": "m2758_0_throatclearing.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12528 + }, + { + "path": "f2172_0_laughter.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12529 + }, + { + "path": "m0557_0_sneeze.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12530 + }, + { + "path": "f2745_0_sniff.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12531 + }, + { + "path": "f2937_0_cough.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12532 + }, + { + "path": "m3325_0_cough.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12533 + }, + { + "path": "f2305_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12534 + }, + { + "path": "f1798_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12535 + }, + { + "path": "m0001_0_laughter.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12536 + }, + { + "path": "f2853_0_cough.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12537 + }, + { + "path": "m1275_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12538 + }, + { + "path": "f2350_0_sigh.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12539 + }, + { + "path": "m0092_0_sigh.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12540 + }, + { + "path": "m1906_0_sigh.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12541 + }, + { + "path": "m1845_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12542 + }, + { + "path": "f3019_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12543 + }, + { + "path": "f2608_0_sigh.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12544 + }, + { + "path": "f0349_0_cough.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12545 + }, + { + "path": "m1938_0_sniff.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12546 + }, + { + "path": "f1941_0_sneeze.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12547 + }, + { + "path": "f2608_0_laughter.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12548 + }, + { + "path": "f1601_0_cough.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12549 + }, + { + "path": "m0282_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12550 + }, + { + "path": "f0217_0_sniff.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12551 + }, + { + "path": "m1382_0_sniff.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12552 + }, + { + "path": "f0631_0_laughter.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12553 + }, + { + "path": "m1916_0_sigh.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12554 + }, + { + "path": "m2037_0_sneeze.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12555 + }, + { + "path": "m2733_1_sniff.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12556 + }, + { + "path": "f1957_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12557 + }, + { + "path": "m1925_0_throatclearing.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12558 + }, + { + "path": "m0932_0_throatclearing.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12559 + }, + { + "path": "f0298_0_sneeze.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12560 + }, + { + "path": "f2644_0_sigh.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12561 + }, + { + "path": "f0909_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12562 + }, + { + "path": "m2200_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12563 + }, + { + "path": "f0238_0_sigh.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12564 + }, + { + "path": "m0844_0_throatclearing.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12565 + }, + { + "path": "f2644_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12566 + }, + { + "path": "f0226_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12567 + }, + { + "path": "m1577_0_sigh.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12568 + }, + { + "path": "f2470_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12569 + }, + { + "path": "m3120_0_cough.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12570 + }, + { + "path": "f1934_0_throatclearing.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12571 + }, + { + "path": "f2388_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12572 + }, + { + "path": "m1275_0_cough.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12573 + }, + { + "path": "m0425_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12574 + }, + { + "path": "f0889_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12575 + }, + { + "path": "m0202_0_sigh.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12576 + }, + { + "path": "f0208_0_sniff.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12577 + }, + { + "path": "m2758_0_sniff.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12578 + }, + { + "path": "f2837_0_sniff.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12579 + }, + { + "path": "m0618_0_laughter.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12580 + }, + { + "path": "m0001_0_sneeze.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12581 + }, + { + "path": "m1208_0_cough.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12582 + }, + { + "path": "f1751_0_sneeze.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12583 + }, + { + "path": "f2436_0_sneeze.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12584 + }, + { + "path": "f2784_0_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12585 + }, + { + "path": "m3292_0_sigh.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12586 + }, + { + "path": "f3053_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12587 + }, + { + "path": "f3323_0_sigh.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12588 + }, + { + "path": "m2976_0_sniff.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12589 + }, + { + "path": "f0638_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12590 + }, + { + "path": "m2401_0_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12591 + }, + { + "path": "m2682_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12592 + }, + { + "path": "f0838_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12593 + }, + { + "path": "f1974_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12594 + }, + { + "path": "f2827_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12595 + }, + { + "path": "m2751_0_sneeze.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12596 + }, + { + "path": "f0593_0_throatclearing.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12597 + }, + { + "path": "m1202_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12598 + }, + { + "path": "f0450_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12599 + }, + { + "path": "f3291_0_sneeze.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12600 + }, + { + "path": "f0128_0_laughter.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12601 + }, + { + "path": "m1905_0_cough.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12602 + }, + { + "path": "m0247_0_laughter.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12603 + }, + { + "path": "m2583_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12604 + }, + { + "path": "f1480_0_sniff.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12605 + }, + { + "path": "m0895_1_laughter.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12606 + }, + { + "path": "f0422_0_sigh.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12607 + }, + { + "path": "m0551_0_sniff.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12608 + }, + { + "path": "m2677_0_throatclearing.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12609 + }, + { + "path": "f1994_0_sneeze.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12610 + }, + { + "path": "f2293_1_laughter.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12611 + }, + { + "path": "f1618_0_sneeze.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12612 + }, + { + "path": "m2851_0_laughter.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12613 + }, + { + "path": "f2293_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12614 + }, + { + "path": "m2999_0_sneeze.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12615 + }, + { + "path": "f0281_0_sneeze.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12616 + }, + { + "path": "m1442_0_laughter.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12617 + }, + { + "path": "f1700_0_sniff.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12618 + }, + { + "path": "m2512_0_sniff.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12619 + }, + { + "path": "m2493_0_sneeze.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12620 + }, + { + "path": "m2954_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12621 + }, + { + "path": "f0101_0_sigh.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12622 + }, + { + "path": "m0113_0_sigh.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12623 + }, + { + "path": "f2459_0_cough.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12624 + }, + { + "path": "m0682_0_laughter.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12625 + }, + { + "path": "f0775_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12626 + }, + { + "path": "m2479_0_cough.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12627 + }, + { + "path": "f1054_0_laughter.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12628 + }, + { + "path": "m1021_0_sneeze.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12629 + }, + { + "path": "m1609_0_sneeze.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12630 + }, + { + "path": "f0181_0_sniff.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12631 + }, + { + "path": "m0895_0_cough.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12632 + }, + { + "path": "m1338_0_sigh.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12633 + }, + { + "path": "m1207_0_laughter.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12634 + }, + { + "path": "m1988_0_sniff.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12635 + }, + { + "path": "m1150_0_sniff.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12636 + }, + { + "path": "m2647_0_laughter.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12637 + }, + { + "path": "f1706_0_laughter.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12638 + }, + { + "path": "m3199_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12639 + }, + { + "path": "m1641_0_throatclearing.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12640 + }, + { + "path": "f2443_0_sniff.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12641 + }, + { + "path": "m2690_0_sniff.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12642 + }, + { + "path": "f0422_0_sniff.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12643 + }, + { + "path": "m0154_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12644 + }, + { + "path": "f2336_0_cough.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12645 + }, + { + "path": "m0026_0_cough.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12646 + }, + { + "path": "f1438_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12647 + }, + { + "path": "m2445_0_sneeze.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12648 + }, + { + "path": "f0471_0_cough.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12649 + }, + { + "path": "f2783_0_laughter.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12650 + }, + { + "path": "f1957_2_sneeze.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12651 + }, + { + "path": "f2478_0_sigh.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12652 + }, + { + "path": "m1426_0_sniff.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12653 + }, + { + "path": "m2081_0_cough.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12654 + }, + { + "path": "m1444_0_sigh.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12655 + }, + { + "path": "f1547_0_sniff.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12656 + }, + { + "path": "f3003_0_sniff.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12657 + }, + { + "path": "m1005_0_sigh.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12658 + }, + { + "path": "o0338_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12659 + }, + { + "path": "m0648_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12660 + }, + { + "path": "m1025_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12661 + }, + { + "path": "f1686_0_sigh.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12662 + }, + { + "path": "m0771_0_throatclearing.wav", + "question": "What's the component that's generating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12663 + }, + { + "path": "m0960_0_throatclearing.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12664 + }, + { + "path": "m0275_0_throatclearing.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12665 + }, + { + "path": "m1351_0_sigh.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12666 + }, + { + "path": "m1180_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12667 + }, + { + "path": "f0585_1_throatclearing.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12668 + }, + { + "path": "f1307_0_sigh.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12669 + }, + { + "path": "m3197_0_sneeze.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12670 + }, + { + "path": "f1801_0_laughter.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12671 + }, + { + "path": "m0595_2_sneeze.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12672 + }, + { + "path": "m3253_0_sneeze.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12673 + }, + { + "path": "m0798_0_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12674 + }, + { + "path": "m0389_0_laughter.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12675 + }, + { + "path": "m0595_1_cough.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12676 + }, + { + "path": "f2043_0_sneeze.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12677 + }, + { + "path": "m0733_0_sniff.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12678 + }, + { + "path": "f0740_0_sigh.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12679 + }, + { + "path": "m2147_0_sneeze.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12680 + }, + { + "path": "f0638_0_sigh.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12681 + }, + { + "path": "f0755_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12682 + }, + { + "path": "f3291_0_sigh.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12683 + }, + { + "path": "f1646_0_laughter.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12684 + }, + { + "path": "f3334_0_sigh.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12685 + }, + { + "path": "f1220_0_sniff.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12686 + }, + { + "path": "m0470_0_throatclearing.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12687 + }, + { + "path": "f1435_0_cough.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12688 + }, + { + "path": "f0926_0_sneeze.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12689 + }, + { + "path": "m3203_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12690 + }, + { + "path": "m1210_0_sneeze.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12691 + }, + { + "path": "m3325_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12692 + }, + { + "path": "f2764_0_throatclearing.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12693 + }, + { + "path": "m0063_0_laughter.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12694 + }, + { + "path": "m1533_0_sigh.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12695 + }, + { + "path": "m1417_0_laughter.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12696 + }, + { + "path": "f0177_0_sneeze.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12697 + }, + { + "path": "f0408_0_laughter.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12698 + }, + { + "path": "m1034_0_throatclearing.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12699 + }, + { + "path": "o1661_0_cough.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12700 + }, + { + "path": "f2958_0_sneeze.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12701 + }, + { + "path": "m1911_0_laughter.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12702 + }, + { + "path": "m2160_0_laughter.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12703 + }, + { + "path": "f1414_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12704 + }, + { + "path": "m2153_0_sigh.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12705 + }, + { + "path": "f1321_0_cough.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12706 + }, + { + "path": "f3149_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12707 + }, + { + "path": "f2350_0_sniff.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12708 + }, + { + "path": "m2964_0_cough.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12709 + }, + { + "path": "f3053_0_sniff.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12710 + }, + { + "path": "m0431_0_sneeze.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12711 + }, + { + "path": "f0165_0_laughter.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12712 + }, + { + "path": "m0799_0_throatclearing.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12713 + }, + { + "path": "f2342_0_throatclearing.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12714 + }, + { + "path": "f1145_0_throatclearing.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12715 + }, + { + "path": "f2779_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12716 + }, + { + "path": "f0217_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12717 + }, + { + "path": "m0817_0_throatclearing.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12718 + }, + { + "path": "m0960_0_sneeze.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12719 + }, + { + "path": "m1609_0_cough.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12720 + }, + { + "path": "f0605_1_sigh.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12721 + }, + { + "path": "m1395_0_sneeze.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12722 + }, + { + "path": "f3027_0_laughter.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12723 + }, + { + "path": "f1007_0_cough.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12724 + }, + { + "path": "m1732_0_throatclearing.wav", + "question": "What is responsible for the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12725 + }, + { + "path": "f0953_0_sigh.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12726 + }, + { + "path": "m0561_1_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12727 + }, + { + "path": "m0846_0_sigh.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12728 + }, + { + "path": "m3015_0_sniff.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12729 + }, + { + "path": "f3332_0_throatclearing.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12730 + }, + { + "path": "m0257_0_cough.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12731 + }, + { + "path": "f1145_0_laughter.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12732 + }, + { + "path": "m0932_0_sigh.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12733 + }, + { + "path": "m1114_0_sigh.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12734 + }, + { + "path": "m2828_0_cough.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12735 + }, + { + "path": "f2187_0_cough.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12736 + }, + { + "path": "f1220_0_laughter.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12737 + }, + { + "path": "f2799_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12738 + }, + { + "path": "m0267_0_laughter.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12739 + }, + { + "path": "m1567_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12740 + }, + { + "path": "f2835_0_sneeze.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12741 + }, + { + "path": "f1878_0_cough.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12742 + }, + { + "path": "m0873_0_sneeze.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12743 + }, + { + "path": "f0101_0_sneeze.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12744 + }, + { + "path": "m1393_0_cough.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12745 + }, + { + "path": "m1306_0_sniff.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12746 + }, + { + "path": "f2248_0_sigh.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12747 + }, + { + "path": "m1735_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12748 + }, + { + "path": "m1426_0_sigh.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12749 + }, + { + "path": "f0298_0_laughter.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12750 + }, + { + "path": "m2037_0_sigh.wav", + "question": "What's the phenomenon causing the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12751 + }, + { + "path": "f2354_0_sniff.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12752 + }, + { + "path": "m2395_0_sigh.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12753 + }, + { + "path": "m2547_0_sneeze.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12754 + }, + { + "path": "f1746_0_sneeze.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12755 + }, + { + "path": "m2395_0_cough.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12756 + }, + { + "path": "m3342_0_throatclearing.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12757 + }, + { + "path": "f0909_0_sigh.wav", + "question": "What could be the source of this sound in the audio?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12758 + }, + { + "path": "m2381_0_cough.wav", + "question": "determine the source of this noise in the audio.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12759 + }, + { + "path": "f2739_0_throatclearing.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12760 + }, + { + "path": "f3222_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12761 + }, + { + "path": "f1385_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12762 + }, + { + "path": "m3109_0_sniff.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12763 + }, + { + "path": "f0929_0_sigh.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12764 + }, + { + "path": "m1382_0_throatclearing.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12765 + }, + { + "path": "f3291_0_cough.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12766 + }, + { + "path": "m0844_0_sneeze.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12767 + }, + { + "path": "m3180_0_throatclearing.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12768 + }, + { + "path": "f1779_0_laughter.wav", + "question": "What's the thing that's emitting the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12769 + }, + { + "path": "m0175_0_sniff.wav", + "question": "figure out the source of the sound in this audio.", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12770 + }, + { + "path": "f1686_0_sneeze.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12771 + }, + { + "path": "f0735_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12772 + }, + { + "path": "m2024_0_sniff.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12773 + }, + { + "path": "f2972_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12774 + }, + { + "path": "f3332_0_sigh.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12775 + }, + { + "path": "m0420_0_sigh.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12776 + }, + { + "path": "m0618_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12777 + }, + { + "path": "m0960_0_sigh.wav", + "question": "pinpoint the origin of this sound in the clip.", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12778 + }, + { + "path": "f1153_0_sigh.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12779 + }, + { + "path": "o1897_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12780 + }, + { + "path": "f0274_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12781 + }, + { + "path": "m1823_0_sniff.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12782 + }, + { + "path": "f2043_0_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12783 + }, + { + "path": "m0223_0_sneeze.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12784 + }, + { + "path": "f2728_0_throatclearing.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12785 + }, + { + "path": "f2106_0_laughter.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Laughter", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12786 + }, + { + "path": "f1467_0_sneeze.wav", + "question": "Where's the sound in this clip originating from?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12787 + }, + { + "path": "f0735_0_laughter.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12788 + }, + { + "path": "f3265_0_laughter.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12789 + }, + { + "path": "m2345_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12790 + }, + { + "path": "m2776_0_cough.wav", + "question": "identify what's behind the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12791 + }, + { + "path": "m1112_0_sniff.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12792 + }, + { + "path": "f2666_0_cough.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12793 + }, + { + "path": "f3069_0_throatclearing.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12794 + }, + { + "path": "f2837_0_cough.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12795 + }, + { + "path": "f3216_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12796 + }, + { + "path": "m1988_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12797 + }, + { + "path": "m0390_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12798 + }, + { + "path": "m2763_0_cough.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12799 + }, + { + "path": "f0410_0_cough.wav", + "question": "What's the originator of the noise in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sniff", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12800 + }, + { + "path": "m2596_0_sigh.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12801 + }, + { + "path": "f2085_0_throatclearing.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12802 + }, + { + "path": "f1177_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12803 + }, + { + "path": "m3013_0_sneeze.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12804 + }, + { + "path": "f0606_0_laughter.wav", + "question": "What's the factor causing the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12805 + }, + { + "path": "m2024_0_sigh.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12806 + }, + { + "path": "m0798_0_throatclearing.wav", + "question": "detect the source of the sound in this audio recording.", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12807 + }, + { + "path": "f0692_0_laughter.wav", + "question": "tell me what's making the sound in this recording.", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12808 + }, + { + "path": "f2898_0_sneeze.wav", + "question": "What's creating the sound in this audio recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12809 + }, + { + "path": "f0756_0_sneeze.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12810 + }, + { + "path": "f2192_0_sigh.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Laughter", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12811 + }, + { + "path": "f1153_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12812 + }, + { + "path": "f0085_0_throatclearing.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12813 + }, + { + "path": "m0325_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12814 + }, + { + "path": "m0665_0_laughter.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12815 + }, + { + "path": "f2649_0_cough.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12816 + }, + { + "path": "f1748_0_sniff.wav", + "question": "Where is the sound in this clip sourced from?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12817 + }, + { + "path": "m1967_0_laughter.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12818 + }, + { + "path": "o2139_0_sigh.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12819 + }, + { + "path": "f2462_0_sigh.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12820 + }, + { + "path": "f2746_0_sneeze.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sneeze", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sniff", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12821 + }, + { + "path": "f3031_0_throatclearing.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12822 + }, + { + "path": "m2372_0_sigh.wav", + "question": "What's the root of the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12823 + }, + { + "path": "f0405_0_sneeze.wav", + "question": "specify what's generating the sound in this recording.", + "choice_a": "Sniff", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12824 + }, + { + "path": "m2335_0_sneeze.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Cough", + "choice_d": "Throat clearing", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12825 + }, + { + "path": "m3250_0_sneeze.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12826 + }, + { + "path": "f3069_0_sniff.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12827 + }, + { + "path": "f1961_0_sniff.wav", + "question": "What's the cause of the sound you hear in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sneeze", + "choice_c": "Sniff", + "choice_d": "Cough", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12828 + }, + { + "path": "m0374_0_sniff.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sniff", + "choice_b": "Sigh", + "choice_c": "Laughter", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12829 + }, + { + "path": "m1927_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12830 + }, + { + "path": "f1438_0_laughter.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12831 + }, + { + "path": "f0424_0_sigh.wav", + "question": "What is the sound in this audio clip coming from?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12832 + }, + { + "path": "f2712_0_throatclearing.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12833 + }, + { + "path": "m1533_0_sniff.wav", + "question": "What's the thing that's causing the sound in this clip?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Throat clearing", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12834 + }, + { + "path": "m1215_0_throatclearing.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12835 + }, + { + "path": "f1639_0_sneeze.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12836 + }, + { + "path": "f1307_0_throatclearing.wav", + "question": "What's the object that's producing the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12837 + }, + { + "path": "m1275_0_sigh.wav", + "question": "What's the provenance of the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12838 + }, + { + "path": "f2750_0_laughter.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12839 + }, + { + "path": "f1156_0_sigh.wav", + "question": "What's the mechanism behind the sound in this audio clip?", + "choice_a": "Laughter", + "choice_b": "Sigh", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12840 + }, + { + "path": "f3143_0_throatclearing.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12841 + }, + { + "path": "f0410_0_sigh.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12842 + }, + { + "path": "m0670_0_sigh.wav", + "question": "What's the contributor to the sound in this audio?", + "choice_a": "Cough", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12843 + }, + { + "path": "f0593_2_throatclearing.wav", + "question": "What's the entity responsible for the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Sneeze", + "choice_d": "Throat clearing", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12844 + }, + { + "path": "m2565_0_sniff.wav", + "question": "What's behind the sound in this audio?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Sniff", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12845 + }, + { + "path": "f0349_0_sneeze.wav", + "question": "What is the source of this sound in this audio clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12846 + }, + { + "path": "m2196_0_sigh.wav", + "question": "What's the sound we're hearing in this audio from?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Sniff", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12847 + }, + { + "path": "m3180_0_cough.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12848 + }, + { + "path": "m1041_0_sniff.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sneeze", + "choice_b": "Sigh", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12849 + }, + { + "path": "m2753_0_throatclearing.wav", + "question": "From what is this sound emanating in the recording?", + "choice_a": "Sigh", + "choice_b": "Sneeze", + "choice_c": "Throat clearing", + "choice_d": "Sniff", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12850 + }, + { + "path": "m2154_0_sigh.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Laughter", + "choice_b": "Sniff", + "choice_c": "Sigh", + "choice_d": "Throat clearing", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12851 + }, + { + "path": "f0388_0_sigh.wav", + "question": "What's the item making the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12852 + }, + { + "path": "m1863_0_throatclearing.wav", + "question": "Where is this sound coming from in the audio?", + "choice_a": "Sigh", + "choice_b": "Throat clearing", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12853 + }, + { + "path": "f0238_0_cough.wav", + "question": "discern the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12854 + }, + { + "path": "f1964_0_sigh.wav", + "question": "What's the instrument of the sound in this clip?", + "choice_a": "Sniff", + "choice_b": "Laughter", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12855 + }, + { + "path": "f0889_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12856 + }, + { + "path": "f1748_0_cough.wav", + "question": "What's the device creating the sound in this recording?", + "choice_a": "Sniff", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12857 + }, + { + "path": "m2803_0_sigh.wav", + "question": "trace the source of the sound in this recording.", + "choice_a": "Throat clearing", + "choice_b": "Sniff", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12858 + }, + { + "path": "f2852_0_laughter.wav", + "question": "What's the machinery that's producing the sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Throat clearing", + "choice_c": "Laughter", + "choice_d": "Sneeze", + "answer_gt": "Laughter", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12859 + }, + { + "path": "f2853_0_sneeze.wav", + "question": "guess the source of the sound in this audio.", + "choice_a": "Sneeze", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12860 + }, + { + "path": "m1444_0_throatclearing.wav", + "question": "What do you think is the source of the sound in this audio?", + "choice_a": "Sniff", + "choice_b": "Throat clearing", + "choice_c": "Sigh", + "choice_d": "Cough", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12861 + }, + { + "path": "m2283_0_cough.wav", + "question": "What's the generator of the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12862 + }, + { + "path": "m1863_0_sniff.wav", + "question": "What's the element that's making the sound in this clip?", + "choice_a": "Cough", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Laughter", + "answer_gt": "Sniff", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12863 + }, + { + "path": "m2976_0_sigh.wav", + "question": "What's the agent producing the sound in this audio?", + "choice_a": "Sigh", + "choice_b": "Laughter", + "choice_c": "Cough", + "choice_d": "Sneeze", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12864 + }, + { + "path": "m1912_0_sigh.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Laughter", + "choice_d": "Sigh", + "answer_gt": "Sigh", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12865 + }, + { + "path": "f1309_0_sneeze.wav", + "question": "What's producing the sound in this recording?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sneeze", + "choice_d": "Sigh", + "answer_gt": "Sneeze", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12866 + }, + { + "path": "f2118_0_throatclearing.wav", + "question": "What's the process that's creating the sound in this clip?", + "choice_a": "Throat clearing", + "choice_b": "Cough", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12867 + }, + { + "path": "m0578_0_cough.wav", + "question": "What's the cause behind the sound in this recording?", + "choice_a": "Sigh", + "choice_b": "Sniff", + "choice_c": "Throat clearing", + "choice_d": "Cough", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12868 + }, + { + "path": "f3249_0_throatclearing.wav", + "question": "What's the material producing the sound in this audio clip?", + "choice_a": "Sneeze", + "choice_b": "Laughter", + "choice_c": "Throat clearing", + "choice_d": "Sigh", + "answer_gt": "Throat clearing", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12869 + }, + { + "path": "f2108_0_cough.wav", + "question": "What's the origin of the audible sound in this audio clip?", + "choice_a": "Cough", + "choice_b": "Sneeze", + "choice_c": "Sigh", + "choice_d": "Laughter", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12870 + }, + { + "path": "m0344_0_cough.wav", + "question": "What's the system behind the sound in this audio?", + "choice_a": "Laughter", + "choice_b": "Cough", + "choice_c": "Sniff", + "choice_d": "Sigh", + "answer_gt": "Cough", + "task_name": "vocal_sound_classification", + "dataset_name": "VocalSound", + "uniq_id": 12871 + }, + { + "path": "Street_user0346_14824996_003.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "kitchen", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "crowded indoor", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12872 + }, + { + "path": "Park_user0323_15518742_002.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "cafe", + "choice_b": "kitchen", + "choice_c": "park", + "choice_d": "subway station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12873 + }, + { + "path": "Car_user0001_14876927_005.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12874 + }, + { + "path": "Kitchen_user0117_14832425_002.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12875 + }, + { + "path": "Subway_user0528_14980295_000.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12876 + }, + { + "path": "Restroom_user0530_14844535_000.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12877 + }, + { + "path": "Restroom_user0248_14842718_000.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12878 + }, + { + "path": "Cafe_user0225_15509118_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "subway", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12879 + }, + { + "path": "Bus_user0704_14884826_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "bus", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12880 + }, + { + "path": "Street_user0718_14831352_005.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12881 + }, + { + "path": "Bus_user0326_14870508_007.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "street", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12882 + }, + { + "path": "Restroom_user0730_14848193_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12883 + }, + { + "path": "Park_user0074_15425576_001.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "street", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12884 + }, + { + "path": "ResidentialArea_user0352_14982448_004.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "crowded indoor", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12885 + }, + { + "path": "Subway_user0200_14875579_005.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "subway station", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12886 + }, + { + "path": "CrowdedIndoor_user0811_14981841_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12887 + }, + { + "path": "SubwayStation_user0267_15053357_001.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12888 + }, + { + "path": "Bus_user0762_14884557_005.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12889 + }, + { + "path": "Park_user0803_14986601_004.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12890 + }, + { + "path": "Restaurant_user0275_15519148_001.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12891 + }, + { + "path": "ResidentialArea_user0817_14984040_003.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12892 + }, + { + "path": "Elevator_user0593_15112839_005.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "kitchen", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "park", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12893 + }, + { + "path": "Restaurant_user0597_14889500_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "car", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12894 + }, + { + "path": "ResidentialArea_user0081_14895443_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12895 + }, + { + "path": "Restaurant_user0817_15020543_005.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "subway", + "choice_b": "restaurant", + "choice_c": "street", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12896 + }, + { + "path": "Subway_user0200_14875591_001.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12897 + }, + { + "path": "SubwayStation_user0389_15519835_006.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12898 + }, + { + "path": "Kitchen_user0658_14836431_003.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "restroom", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12899 + }, + { + "path": "CrowdedIndoor_user0461_14894935_005.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "car", + "choice_b": "restroom", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12900 + }, + { + "path": "Subway_user0435_14876215_005.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12901 + }, + { + "path": "Cafe_user0266_14977244_004.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "car", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "residential area", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12902 + }, + { + "path": "Cafe_user0066_14986521_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12903 + }, + { + "path": "SubwayStation_user0698_15507908_001.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12904 + }, + { + "path": "CrowdedIndoor_user0593_14984852_001.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway station", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12905 + }, + { + "path": "Elevator_user0009_14987113_002.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12906 + }, + { + "path": "Park_user0533_15119089_000.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12907 + }, + { + "path": "Bus_user0630_14979553_004.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12908 + }, + { + "path": "Street_user0493_14821515_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "street", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "kitchen", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12909 + }, + { + "path": "SubwayStation_user0378_14985982_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12910 + }, + { + "path": "ResidentialArea_user0243_14889996_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12911 + }, + { + "path": "Car_user0479_14873121_003.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12912 + }, + { + "path": "Restaurant_user0755_14879459_006.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12913 + }, + { + "path": "SubwayStation_user0277_14880134_003.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12914 + }, + { + "path": "Restroom_user0667_14818305_003.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12915 + }, + { + "path": "Street_user0250_14827294_003.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "park", + "choice_d": "restaurant", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12916 + }, + { + "path": "Park_user0775_14983643_002.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12917 + }, + { + "path": "Restaurant_user0652_14981631_003.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "crowded indoor", + "choice_b": "restaurant", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12918 + }, + { + "path": "Car_user0348_14881130_005.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12919 + }, + { + "path": "CrowdedIndoor_user0811_14981739_002.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "subway", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12920 + }, + { + "path": "ResidentialArea_user0560_14980547_005.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12921 + }, + { + "path": "Restroom_user0368_14821323_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12922 + }, + { + "path": "CrowdedIndoor_user0057_14870194_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12923 + }, + { + "path": "Restroom_user0417_14843645_003.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12924 + }, + { + "path": "Elevator_user0596_15311989_003.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12925 + }, + { + "path": "Bus_user0200_14875499_002.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "crowded indoor", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12926 + }, + { + "path": "Elevator_user0822_15000601_001.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12927 + }, + { + "path": "Cafe_user0027_15441285_002.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12928 + }, + { + "path": "Park_user0526_15119568_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12929 + }, + { + "path": "Kitchen_user0599_14822314_002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12930 + }, + { + "path": "Restroom_user0314_14821320_005.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "crowded indoor", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12931 + }, + { + "path": "CrowdedIndoor_user0461_14894905_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restaurant", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12932 + }, + { + "path": "Park_user0274_15053855_005.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12933 + }, + { + "path": "Cafe_user0802_14892056_005.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "car", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12934 + }, + { + "path": "Bus_user0014_14981988_003.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "cafe", + "choice_b": "elevator", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12935 + }, + { + "path": "ResidentialArea_user0276_14868981_000.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12936 + }, + { + "path": "Bus_user0224_14980190_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12937 + }, + { + "path": "Cafe_user0027_15517570_002.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12938 + }, + { + "path": "Bus_user0365_14982374_004.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12939 + }, + { + "path": "Restaurant_user0132_14988093_000.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12940 + }, + { + "path": "Car_user0007_14855563_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12941 + }, + { + "path": "CrowdedIndoor_user0346_14888094_000.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "car", + "choice_b": "crowded indoor", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12942 + }, + { + "path": "Elevator_user0221_14981232_004.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12943 + }, + { + "path": "SubwayStation_user0277_14880115_005.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12944 + }, + { + "path": "SubwayStation_user0698_15508122_004.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "cafe", + "choice_b": "elevator", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12945 + }, + { + "path": "Park_user0769_14981009_003.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12946 + }, + { + "path": "Restaurant_user0089_15053655_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12947 + }, + { + "path": "Subway_user0151_14879878_001.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "kitchen", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12948 + }, + { + "path": "Restroom_user0301_14828230_004.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restroom", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12949 + }, + { + "path": "Car_user0423_14861203_003.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12950 + }, + { + "path": "Bus_user0653_14869731_002.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "car", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12951 + }, + { + "path": "CrowdedIndoor_user0105_14871571_004.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12952 + }, + { + "path": "Elevator_user0122_15053445_001.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12953 + }, + { + "path": "SubwayStation_user0138_14984630_016.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "restroom", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12954 + }, + { + "path": "Restroom_user0013_14833379_004.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12955 + }, + { + "path": "Cafe_user0620_14894692_004.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12956 + }, + { + "path": "Subway_user0148_14897595_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "elevator", + "choice_d": "crowded indoor", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12957 + }, + { + "path": "Restroom_user0618_14849891_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12958 + }, + { + "path": "Bus_user0239_14868208_003.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12959 + }, + { + "path": "Bus_user0704_14884726_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "cafe", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12960 + }, + { + "path": "CrowdedIndoor_user0668_14880209_001.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "crowded indoor", + "choice_b": "elevator", + "choice_c": "bus", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12961 + }, + { + "path": "SubwayStation_user0118_14886512_003.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12962 + }, + { + "path": "Street_user0769_14833514_000.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12963 + }, + { + "path": "SubwayStation_user0586_14888949_005.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12964 + }, + { + "path": "SubwayStation_user0403_15508250_003.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12965 + }, + { + "path": "Car_user0692_14853357_000.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12966 + }, + { + "path": "Car_user0126_14848815_000.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12967 + }, + { + "path": "SubwayStation_user0381_15208819_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12968 + }, + { + "path": "SubwayStation_user0378_14985982_004.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12969 + }, + { + "path": "Restaurant_user0799_14981029_002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12970 + }, + { + "path": "Bus_user0200_14875447_003.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "kitchen", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12971 + }, + { + "path": "CrowdedIndoor_user0159_14875904_004.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "subway", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12972 + }, + { + "path": "Cafe_user0598_15518986_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12973 + }, + { + "path": "Elevator_user0239_14986841_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12974 + }, + { + "path": "Street_user0448_14852420_005.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12975 + }, + { + "path": "Restaurant_user0500_15518827_005.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "crowded indoor", + "choice_b": "restaurant", + "choice_c": "subway", + "choice_d": "subway station", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12976 + }, + { + "path": "Bus_user0704_14884826_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12977 + }, + { + "path": "Elevator_user0306_14896474_004.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12978 + }, + { + "path": "Park_user0698_15508524_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "kitchen", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12979 + }, + { + "path": "SubwayStation_user0267_15053357_002.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12980 + }, + { + "path": "Park_user0657_15235517_000.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "street", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12981 + }, + { + "path": "Kitchen_user0811_14854026_002.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "restaurant", + "choice_d": "restroom", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12982 + }, + { + "path": "Restroom_user0587_14818937_002.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12983 + }, + { + "path": "Elevator_user0801_15125894_005.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "restaurant", + "choice_b": "crowded indoor", + "choice_c": "bus", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12984 + }, + { + "path": "Cafe_user0433_15311922_000.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12985 + }, + { + "path": "Restaurant_user0089_15440865_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12986 + }, + { + "path": "Restroom_user0692_14840741_004.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12987 + }, + { + "path": "Subway_user0148_14897664_002.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12988 + }, + { + "path": "Kitchen_user0425_14854091_003.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "subway", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12989 + }, + { + "path": "Elevator_user0593_15112804_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12990 + }, + { + "path": "Subway_user0148_14897688_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12991 + }, + { + "path": "Cafe_user0216_14981794_000.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12992 + }, + { + "path": "Restaurant_user0099_15122216_005.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "subway", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12993 + }, + { + "path": "Cafe_user0795_14987030_002.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "street", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12994 + }, + { + "path": "Bus_user0658_14869138_003.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12995 + }, + { + "path": "Street_user0769_14833417_003.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "kitchen", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12996 + }, + { + "path": "Subway_user0065_14872486_002.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "subway station", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12997 + }, + { + "path": "Elevator_user0398_15441173_006.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12998 + }, + { + "path": "Restaurant_user0240_15519010_004.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 12999 + }, + { + "path": "SubwayStation_user0829_14893713_000.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "restroom", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13000 + }, + { + "path": "Park_user0173_15519005_005.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13001 + }, + { + "path": "SubwayStation_user0403_15508250_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13002 + }, + { + "path": "Elevator_user0334_15518928_003.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13003 + }, + { + "path": "ResidentialArea_user0536_14985905_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13004 + }, + { + "path": "Car_user0075_14852509_001.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13005 + }, + { + "path": "Car_user0070_14855308_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13006 + }, + { + "path": "Car_user0056_14874961_001.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13007 + }, + { + "path": "Bus_user0138_14982391_006.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13008 + }, + { + "path": "CrowdedIndoor_user0670_14877027_002.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13009 + }, + { + "path": "SubwayStation_user0357_15119501_005.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13010 + }, + { + "path": "Subway_user0285_14979210_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "street", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13011 + }, + { + "path": "Restaurant_user0066_15311055_005.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "subway", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13012 + }, + { + "path": "Car_user0213_14877842_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13013 + }, + { + "path": "Bus_user0704_14884726_004.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13014 + }, + { + "path": "Restroom_user0735_14836952_005.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "residential area", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13015 + }, + { + "path": "Car_user0070_14855337_003.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13016 + }, + { + "path": "SubwayStation_user0378_14986009_001.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13017 + }, + { + "path": "CrowdedIndoor_user0811_14981731_003.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13018 + }, + { + "path": "Park_user0536_14985881_003.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13019 + }, + { + "path": "Restroom_user0343_14829654_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13020 + }, + { + "path": "Restaurant_user0626_15311298_003.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13021 + }, + { + "path": "Kitchen_user0229_14863840_001.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "car", + "choice_b": "crowded indoor", + "choice_c": "kitchen", + "choice_d": "subway station", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13022 + }, + { + "path": "Park_user0323_15518740_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13023 + }, + { + "path": "Subway_user0528_14980298_005.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "bus", + "choice_b": "restaurant", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13024 + }, + { + "path": "Restaurant_user0567_14894330_002.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13025 + }, + { + "path": "Restaurant_user0807_14884516_001.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "restroom", + "choice_c": "restaurant", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13026 + }, + { + "path": "Kitchen_user0316_14850326_003.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13027 + }, + { + "path": "Park_user0266_15119076_000.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13028 + }, + { + "path": "Car_user0453_14882560_004.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13029 + }, + { + "path": "SubwayStation_user0290_15519810_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13030 + }, + { + "path": "Cafe_user0057_14977311_000.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13031 + }, + { + "path": "Restaurant_user0755_14897372_003.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "subway", + "choice_b": "restaurant", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13032 + }, + { + "path": "Park_user0498_15520217_000.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13033 + }, + { + "path": "Restaurant_user0384_14977863_005.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "kitchen", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13034 + }, + { + "path": "Subway_user0799_14979597_005.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "street", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13035 + }, + { + "path": "Street_user0069_14824730_000.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13036 + }, + { + "path": "Cafe_user0718_15518853_002.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13037 + }, + { + "path": "Kitchen_user0658_14836362_000.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13038 + }, + { + "path": "Restroom_user0587_14819025_001.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13039 + }, + { + "path": "ResidentialArea_user0108_14867436_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "residential area", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13040 + }, + { + "path": "Car_user0220_14876410_002.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "elevator", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13041 + }, + { + "path": "Bus_user0762_14884551_003.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "bus", + "choice_b": "subway", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13042 + }, + { + "path": "CrowdedIndoor_user0159_14894795_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13043 + }, + { + "path": "Bus_user0704_14884726_005.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13044 + }, + { + "path": "Bus_user0017_14896159_003.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "park", + "choice_b": "crowded indoor", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13045 + }, + { + "path": "Elevator_user0250_14985686_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13046 + }, + { + "path": "Park_user0741_15518846_005.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13047 + }, + { + "path": "Street_user0106_14827671_005.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13048 + }, + { + "path": "SubwayStation_user0403_15508121_004.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13049 + }, + { + "path": "Cafe_user0626_15311285_002.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13050 + }, + { + "path": "ResidentialArea_user0560_14980214_002.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13051 + }, + { + "path": "CrowdedIndoor_user0595_14889726_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "car", + "choice_b": "crowded indoor", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13052 + }, + { + "path": "Subway_user0109_14890379_004.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13053 + }, + { + "path": "Elevator_user0009_14984656_001.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway station", + "choice_b": "park", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13054 + }, + { + "path": "ResidentialArea_user0567_14868353_000.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "crowded indoor", + "choice_d": "subway station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13055 + }, + { + "path": "Park_user0651_14978862_001.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13056 + }, + { + "path": "CrowdedIndoor_user0461_14894894_002.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "kitchen", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13057 + }, + { + "path": "Cafe_user0433_15311929_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13058 + }, + { + "path": "SubwayStation_user0044_15441278_000.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13059 + }, + { + "path": "Restaurant_user0334_15519023_001.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13060 + }, + { + "path": "CrowdedIndoor_user0188_14863794_003.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13061 + }, + { + "path": "Street_user0183_14830598_000.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "bus", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13062 + }, + { + "path": "ResidentialArea_user0755_14884555_000.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13063 + }, + { + "path": "Subway_user0435_14876275_001.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13064 + }, + { + "path": "Kitchen_user0049_14834868_005.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "street", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13065 + }, + { + "path": "Park_user0651_14978862_000.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13066 + }, + { + "path": "Park_user0498_15520217_003.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13067 + }, + { + "path": "Street_user0106_14827711_000.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13068 + }, + { + "path": "Elevator_user0593_15112804_003.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13069 + }, + { + "path": "ResidentialArea_user0479_14869535_003.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13070 + }, + { + "path": "CrowdedIndoor_user0453_14882407_004.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "street", + "choice_b": "subway", + "choice_c": "cafe", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13071 + }, + { + "path": "Restroom_user0422_14839921_001.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13072 + }, + { + "path": "Restroom_user0417_14843636_005.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13073 + }, + { + "path": "Kitchen_user0445_14861792_000.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "subway", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13074 + }, + { + "path": "Subway_user0200_14875655_005.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "subway", + "choice_d": "subway station", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13075 + }, + { + "path": "Restroom_user0618_14849927_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13076 + }, + { + "path": "Park_user0657_15235517_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13077 + }, + { + "path": "Kitchen_user0445_14861774_001.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13078 + }, + { + "path": "Restaurant_user0626_15311294_002.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13079 + }, + { + "path": "Cafe_user0025_15441137_003.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13080 + }, + { + "path": "Cafe_user0570_15359500_005.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "restroom", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13081 + }, + { + "path": "CrowdedIndoor_user0105_14871624_002.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13082 + }, + { + "path": "Cafe_user0225_15512292_006.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13083 + }, + { + "path": "CrowdedIndoor_user0585_14879918_002.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13084 + }, + { + "path": "Kitchen_user0592_14861952_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13085 + }, + { + "path": "Bus_user0630_14980872_001.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13086 + }, + { + "path": "Car_user0692_14853353_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "crowded indoor", + "choice_b": "elevator", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13087 + }, + { + "path": "SubwayStation_user0091_15518744_000.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "restroom", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13088 + }, + { + "path": "Street_user0254_14823497_005.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13089 + }, + { + "path": "Street_user0212_14828151_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13090 + }, + { + "path": "Car_user0725_14884118_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "subway", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13091 + }, + { + "path": "Kitchen_user0314_14818063_004.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13092 + }, + { + "path": "CrowdedIndoor_user0667_14867924_003.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13093 + }, + { + "path": "Park_user0323_15518738_002.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13094 + }, + { + "path": "Kitchen_user0453_14831257_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13095 + }, + { + "path": "SubwayStation_user0799_15013024_000.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "crowded indoor", + "choice_d": "subway", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13096 + }, + { + "path": "CrowdedIndoor_user0605_14983876_004.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "street", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13097 + }, + { + "path": "Restroom_user0789_14825722_004.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restroom", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13098 + }, + { + "path": "SubwayStation_user0226_14980893_002.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe", + "choice_b": "residential area", + "choice_c": "restaurant", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13099 + }, + { + "path": "Restaurant_user0626_15311292_002.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "restaurant", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13100 + }, + { + "path": "Car_user0075_14852536_002.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13101 + }, + { + "path": "SubwayStation_user0045_15119578_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13102 + }, + { + "path": "CrowdedIndoor_user0479_14875993_001.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13103 + }, + { + "path": "CrowdedIndoor_user0518_14983642_001.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "cafe", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13104 + }, + { + "path": "Bus_user0200_14875469_003.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13105 + }, + { + "path": "Cafe_user0581_14983118_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13106 + }, + { + "path": "Bus_user0408_14869906_001.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13107 + }, + { + "path": "Subway_user0293_14878997_003.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13108 + }, + { + "path": "Car_user0735_14852112_005.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "subway station", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13109 + }, + { + "path": "CrowdedIndoor_user0049_14880152_001.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway station", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13110 + }, + { + "path": "SubwayStation_user0138_14984630_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13111 + }, + { + "path": "Subway_user0468_14869824_000.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "restaurant", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13112 + }, + { + "path": "Park_user0226_14982492_005.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13113 + }, + { + "path": "Elevator_user0074_15430303_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13114 + }, + { + "path": "SubwayStation_user0542_15112838_002.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "cafe", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13115 + }, + { + "path": "Park_user0448_14890871_001.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13116 + }, + { + "path": "Bus_user0762_14885404_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13117 + }, + { + "path": "ResidentialArea_user0580_14884604_005.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "car", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13118 + }, + { + "path": "Cafe_user0718_15518858_001.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13119 + }, + { + "path": "Subway_user0491_14869748_004.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13120 + }, + { + "path": "Park_user0352_14982354_002.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13121 + }, + { + "path": "Park_user0623_14894892_003.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "park", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13122 + }, + { + "path": "Bus_user0326_14980081_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13123 + }, + { + "path": "SubwayStation_user0817_15054416_005.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13124 + }, + { + "path": "Restaurant_user0762_14884694_001.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13125 + }, + { + "path": "Bus_user0448_14879317_001.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "subway station", + "choice_b": "cafe", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13126 + }, + { + "path": "Car_user0220_14876427_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13127 + }, + { + "path": "Bus_user0760_14891487_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13128 + }, + { + "path": "Park_user0016_14983321_005.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "kitchen", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13129 + }, + { + "path": "Restaurant_user0334_15519025_000.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "elevator", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13130 + }, + { + "path": "Bus_user0138_14982370_004.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "cafe", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13131 + }, + { + "path": "Cafe_user0307_15518749_000.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13132 + }, + { + "path": "Street_user0183_14830290_005.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13133 + }, + { + "path": "Elevator_user0680_14983884_001.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "cafe", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13134 + }, + { + "path": "Car_user0797_14875819_003.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "car", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13135 + }, + { + "path": "Cafe_user0346_14978601_002.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13136 + }, + { + "path": "Car_user0015_14853405_005.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "restroom", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13137 + }, + { + "path": "Street_user0314_14822378_004.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13138 + }, + { + "path": "Bus_user0704_14884773_003.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "subway", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13139 + }, + { + "path": "Bus_user0652_14880602_001.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13140 + }, + { + "path": "Subway_user0694_14872850_000.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "park", + "choice_d": "subway station", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13141 + }, + { + "path": "CrowdedIndoor_user0001_14876863_000.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13142 + }, + { + "path": "ResidentialArea_user0580_14884600_000.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13143 + }, + { + "path": "Kitchen_user0177_14857689_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "restaurant", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13144 + }, + { + "path": "Subway_user0591_14868894_002.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "residential area", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13145 + }, + { + "path": "Street_user0465_14834738_004.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13146 + }, + { + "path": "Bus_user0675_14869515_003.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13147 + }, + { + "path": "Restroom_user0049_14837132_004.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "restroom", + "choice_d": "crowded indoor", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13148 + }, + { + "path": "Cafe_user0500_14987716_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13149 + }, + { + "path": "ResidentialArea_user0828_14983581_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "residential area", + "choice_d": "street", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13150 + }, + { + "path": "Subway_user0065_14870435_006.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13151 + }, + { + "path": "Restroom_user0735_14836952_003.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13152 + }, + { + "path": "Street_user0408_14826629_003.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13153 + }, + { + "path": "Car_user0453_14882534_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "subway station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13154 + }, + { + "path": "Park_user0352_14982354_004.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13155 + }, + { + "path": "Street_user0465_14834748_004.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "restaurant", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13156 + }, + { + "path": "Park_user0380_15519726_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13157 + }, + { + "path": "Bus_user0138_14982403_007.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "residential area", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13158 + }, + { + "path": "Subway_user0293_14880859_003.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "park", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13159 + }, + { + "path": "SubwayStation_user0799_15011492_002.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13160 + }, + { + "path": "Park_user0560_14979261_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13161 + }, + { + "path": "Restroom_user0031_14826470_002.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13162 + }, + { + "path": "Kitchen_user0453_14831281_004.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13163 + }, + { + "path": "Street_user0712_14832917_005.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "street", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13164 + }, + { + "path": "SubwayStation_user0226_14980893_000.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13165 + }, + { + "path": "Restroom_user0036_14830609_005.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13166 + }, + { + "path": "Park_user0381_15208821_004.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13167 + }, + { + "path": "Restaurant_user0045_15053600_002.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "bus", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13168 + }, + { + "path": "Restroom_user0386_14849434_005.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "subway station", + "choice_b": "subway", + "choice_c": "residential area", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13169 + }, + { + "path": "Elevator_user0132_15509086_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "kitchen", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13170 + }, + { + "path": "ResidentialArea_user0680_14983897_001.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "crowded indoor", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13171 + }, + { + "path": "Subway_user0318_14875379_004.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13172 + }, + { + "path": "Car_user0285_14855440_001.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13173 + }, + { + "path": "SubwayStation_user0000_15245944_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway station", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13174 + }, + { + "path": "CrowdedIndoor_user0670_14877010_003.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13175 + }, + { + "path": "Street_user0789_14841175_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "crowded indoor", + "choice_b": "restroom", + "choice_c": "subway", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13176 + }, + { + "path": "Car_user0285_14855315_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13177 + }, + { + "path": "Street_user0597_14823227_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13178 + }, + { + "path": "ResidentialArea_user0024_14866966_001.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13179 + }, + { + "path": "Street_user0465_14834748_003.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "street", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13180 + }, + { + "path": "Restroom_user0365_14830208_003.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13181 + }, + { + "path": "SubwayStation_user0378_14985979_005.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "restaurant", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13182 + }, + { + "path": "ResidentialArea_user0421_14986628_002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13183 + }, + { + "path": "Subway_user0820_14892881_000.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13184 + }, + { + "path": "Subway_user0109_14890467_005.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13185 + }, + { + "path": "Car_user0220_14876419_000.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "crowded indoor", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13186 + }, + { + "path": "Restroom_user0365_14830208_000.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "restroom", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13187 + }, + { + "path": "Kitchen_user0386_14845646_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "street", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13188 + }, + { + "path": "Park_user0698_15508525_000.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "subway", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13189 + }, + { + "path": "CrowdedIndoor_user0159_14875904_000.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13190 + }, + { + "path": "ResidentialArea_user0597_14982264_001.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13191 + }, + { + "path": "Restaurant_user0584_15016556_001.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13192 + }, + { + "path": "SubwayStation_user0698_15507908_006.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13193 + }, + { + "path": "Bus_user0658_14869138_004.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13194 + }, + { + "path": "CrowdedIndoor_user0216_14871414_005.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13195 + }, + { + "path": "Subway_user0148_14897595_001.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "restroom", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13196 + }, + { + "path": "Restroom_user0426_14818591_004.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13197 + }, + { + "path": "Cafe_user0216_14981779_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "subway station", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13198 + }, + { + "path": "Subway_user0485_14866784_005.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13199 + }, + { + "path": "CrowdedIndoor_user0057_14870194_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13200 + }, + { + "path": "Cafe_user0378_14895776_005.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "subway", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13201 + }, + { + "path": "Park_user0807_14975811_003.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13202 + }, + { + "path": "Cafe_user0057_14980612_008.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "elevator", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13203 + }, + { + "path": "Cafe_user0581_14899748_001.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "elevator", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13204 + }, + { + "path": "Elevator_user0173_15518949_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13205 + }, + { + "path": "Subway_user0118_14886486_005.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13206 + }, + { + "path": "CrowdedIndoor_user0201_14870094_002.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13207 + }, + { + "path": "Restroom_user0140_14830538_000.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "cafe", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13208 + }, + { + "path": "Car_user0348_14881126_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13209 + }, + { + "path": "Bus_user0704_14884773_000.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13210 + }, + { + "path": "Park_user0623_14894898_004.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13211 + }, + { + "path": "ResidentialArea_user0081_14875972_004.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13212 + }, + { + "path": "Restaurant_user0530_14983127_003.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13213 + }, + { + "path": "ResidentialArea_user0628_14983028_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "restroom", + "choice_b": "cafe", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13214 + }, + { + "path": "Cafe_user0570_15359497_002.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13215 + }, + { + "path": "Elevator_user0822_15000601_000.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13216 + }, + { + "path": "Park_user0596_15119596_003.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13217 + }, + { + "path": "ResidentialArea_user0075_14862264_005.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13218 + }, + { + "path": "Restroom_user0453_14818486_001.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13219 + }, + { + "path": "ResidentialArea_user0777_14983231_001.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "crowded indoor", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13220 + }, + { + "path": "CrowdedIndoor_user0384_14977856_004.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13221 + }, + { + "path": "CrowdedIndoor_user0601_14895635_004.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "crowded indoor", + "choice_b": "restaurant", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13222 + }, + { + "path": "ResidentialArea_user0398_15017725_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "restroom", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13223 + }, + { + "path": "Bus_user0365_14982382_002.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13224 + }, + { + "path": "Car_user0056_14874978_003.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13225 + }, + { + "path": "Park_user0498_15119291_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "subway station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13226 + }, + { + "path": "Restroom_user0750_14816248_008.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13227 + }, + { + "path": "Restroom_user0408_14817928_004.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13228 + }, + { + "path": "Subway_user0118_14886499_001.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13229 + }, + { + "path": "Street_user0575_14833204_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "residential area", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13230 + }, + { + "path": "Park_user0067_15519727_001.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13231 + }, + { + "path": "ResidentialArea_user0500_14867583_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13232 + }, + { + "path": "Cafe_user0159_14987436_005.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13233 + }, + { + "path": "Restroom_user0194_14839029_004.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13234 + }, + { + "path": "Kitchen_user0404_14864393_000.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "cafe", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13235 + }, + { + "path": "Elevator_user0699_14880725_005.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13236 + }, + { + "path": "ResidentialArea_user0108_14867461_004.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13237 + }, + { + "path": "Bus_user0630_14979554_000.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13238 + }, + { + "path": "Subway_user0118_14886499_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13239 + }, + { + "path": "Cafe_user0027_15441317_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "restroom", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13240 + }, + { + "path": "Elevator_user0476_14986657_005.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13241 + }, + { + "path": "Kitchen_user0404_14864402_002.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13242 + }, + { + "path": "Restaurant_user0533_15119130_004.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13243 + }, + { + "path": "Park_user0657_15235773_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13244 + }, + { + "path": "Bus_user0595_14894180_004.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13245 + }, + { + "path": "Cafe_user0609_14888302_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "restroom", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13246 + }, + { + "path": "Kitchen_user0223_14834671_004.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "street", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13247 + }, + { + "path": "Restaurant_user0066_15311059_001.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13248 + }, + { + "path": "ResidentialArea_user0684_14977500_001.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13249 + }, + { + "path": "Street_user0314_14822457_002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13250 + }, + { + "path": "Kitchen_user0131_14841386_004.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13251 + }, + { + "path": "Park_user0352_15053734_001.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "subway", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13252 + }, + { + "path": "CrowdedIndoor_user0290_14867589_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "crowded indoor", + "choice_b": "elevator", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13253 + }, + { + "path": "Subway_user0118_14886486_002.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "street", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13254 + }, + { + "path": "Subway_user0694_14872907_001.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13255 + }, + { + "path": "Park_user0769_14981009_005.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "park", + "choice_b": "subway", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13256 + }, + { + "path": "Restaurant_user0334_15519022_005.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13257 + }, + { + "path": "Street_user0183_14830598_003.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "cafe", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13258 + }, + { + "path": "ResidentialArea_user0536_14985908_003.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13259 + }, + { + "path": "Park_user0014_15016451_002.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "park", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13260 + }, + { + "path": "Park_user0596_15440296_000.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "park", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13261 + }, + { + "path": "SubwayStation_user0799_14979621_003.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13262 + }, + { + "path": "Restroom_user0667_14823232_001.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13263 + }, + { + "path": "Subway_user0820_14892881_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13264 + }, + { + "path": "SubwayStation_user0290_15519810_001.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13265 + }, + { + "path": "Bus_user0326_14870508_008.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13266 + }, + { + "path": "Kitchen_user0365_14829942_004.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13267 + }, + { + "path": "Restroom_user0051_14820823_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13268 + }, + { + "path": "Restroom_user0489_14848544_005.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "crowded indoor", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13269 + }, + { + "path": "Car_user0001_14876892_002.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13270 + }, + { + "path": "Cafe_user0159_14987435_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13271 + }, + { + "path": "Restaurant_user0755_14897360_004.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13272 + }, + { + "path": "CrowdedIndoor_user0365_14981945_003.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13273 + }, + { + "path": "SubwayStation_user0291_14885689_003.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "subway station", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13274 + }, + { + "path": "Restroom_user0298_14846576_004.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13275 + }, + { + "path": "Elevator_user0596_15119386_001.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "cafe", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13276 + }, + { + "path": "Park_user0803_14986601_002.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13277 + }, + { + "path": "Elevator_user0458_15487740_004.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "elevator", + "choice_d": "street", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13278 + }, + { + "path": "Restroom_user0791_14816953_000.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "kitchen", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13279 + }, + { + "path": "Elevator_user0493_14911587_001.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "subway station", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13280 + }, + { + "path": "Park_user0526_15520179_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13281 + }, + { + "path": "Restroom_user0117_14833326_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "cafe", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13282 + }, + { + "path": "Kitchen_user0500_14826615_002.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "street", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "cafe", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13283 + }, + { + "path": "Elevator_user0552_15518841_005.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13284 + }, + { + "path": "Park_user0274_15053860_002.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "cafe", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13285 + }, + { + "path": "Restaurant_user0665_14881053_004.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "bus", + "choice_b": "street", + "choice_c": "restroom", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13286 + }, + { + "path": "Bus_user0276_14877616_000.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13287 + }, + { + "path": "SubwayStation_user0325_15030729_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13288 + }, + { + "path": "Kitchen_user0087_14833019_006.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13289 + }, + { + "path": "Restroom_user0587_14818937_000.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13290 + }, + { + "path": "Car_user0725_14884107_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13291 + }, + { + "path": "Cafe_user0433_15311925_001.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13292 + }, + { + "path": "Kitchen_user0606_14845978_002.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13293 + }, + { + "path": "Park_user0066_14986531_000.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13294 + }, + { + "path": "Restaurant_user0334_15519024_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13295 + }, + { + "path": "Bus_user0675_14979949_000.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13296 + }, + { + "path": "Cafe_user0460_15519442_000.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13297 + }, + { + "path": "Subway_user0468_14869824_001.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "bus", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13298 + }, + { + "path": "Restroom_user0640_14818814_001.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13299 + }, + { + "path": "SubwayStation_user0148_14922311_005.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13300 + }, + { + "path": "SubwayStation_user0528_14980285_002.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13301 + }, + { + "path": "SubwayStation_user0403_15441334_001.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway station", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13302 + }, + { + "path": "SubwayStation_user0528_14980334_002.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13303 + }, + { + "path": "Kitchen_user0805_14840461_001.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13304 + }, + { + "path": "Restaurant_user0498_15311800_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "park", + "choice_b": "subway", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13305 + }, + { + "path": "Restaurant_user0807_14884481_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "restaurant", + "choice_d": "residential area", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13306 + }, + { + "path": "Subway_user0694_14872850_005.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "restroom", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13307 + }, + { + "path": "Restroom_user0781_14837074_000.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13308 + }, + { + "path": "SubwayStation_user0148_14898228_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13309 + }, + { + "path": "SubwayStation_user0017_15311283_001.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13310 + }, + { + "path": "SubwayStation_user0799_15013024_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13311 + }, + { + "path": "CrowdedIndoor_user0600_14986034_002.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "park", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13312 + }, + { + "path": "Subway_user0318_14875405_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "subway", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13313 + }, + { + "path": "Elevator_user0277_15306452_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "crowded indoor", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13314 + }, + { + "path": "CrowdedIndoor_user0811_14981740_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "subway", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13315 + }, + { + "path": "Bus_user0589_14871577_004.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "crowded indoor", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "subway", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13316 + }, + { + "path": "Kitchen_user0404_14864445_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13317 + }, + { + "path": "Park_user0807_14973602_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13318 + }, + { + "path": "Restroom_user0453_14818486_004.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "car", + "choice_b": "elevator", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13319 + }, + { + "path": "SubwayStation_user0528_14980290_005.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "street", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13320 + }, + { + "path": "Restroom_user0031_14816436_005.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13321 + }, + { + "path": "Kitchen_user0667_14823379_000.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "car", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13322 + }, + { + "path": "CrowdedIndoor_user0015_14870152_004.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13323 + }, + { + "path": "ResidentialArea_user0828_14983583_001.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "street", + "choice_b": "elevator", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13324 + }, + { + "path": "Elevator_user0099_15113474_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13325 + }, + { + "path": "Restroom_user0692_14840693_002.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13326 + }, + { + "path": "SubwayStation_user0000_15245944_000.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13327 + }, + { + "path": "Street_user0369_14831719_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "street", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13328 + }, + { + "path": "SubwayStation_user0438_15119086_007.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13329 + }, + { + "path": "Park_user0803_14986596_005.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13330 + }, + { + "path": "SubwayStation_user0357_15441268_003.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13331 + }, + { + "path": "Bus_user0653_14869727_005.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "subway station", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13332 + }, + { + "path": "Bus_user0200_14875469_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "restaurant", + "choice_d": "elevator", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13333 + }, + { + "path": "Park_user0323_15518742_005.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13334 + }, + { + "path": "Car_user0095_14881310_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13335 + }, + { + "path": "Restroom_user0659_14823378_000.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13336 + }, + { + "path": "Restaurant_user0151_14879931_004.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "street", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13337 + }, + { + "path": "Kitchen_user0772_14855770_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "bus", + "choice_b": "kitchen", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13338 + }, + { + "path": "Restroom_user0606_14823344_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13339 + }, + { + "path": "CrowdedIndoor_user0653_14985863_004.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "elevator", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13340 + }, + { + "path": "Kitchen_user0453_14831281_001.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "kitchen", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13341 + }, + { + "path": "ResidentialArea_user0421_14986629_004.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13342 + }, + { + "path": "Restaurant_user0500_15518826_002.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "restaurant", + "choice_d": "crowded indoor", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13343 + }, + { + "path": "Kitchen_user0108_14849313_002.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13344 + }, + { + "path": "Elevator_user0334_15518928_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13345 + }, + { + "path": "Car_user0575_14866590_001.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "car", + "choice_d": "cafe", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13346 + }, + { + "path": "Kitchen_user0365_14829905_000.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "subway", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13347 + }, + { + "path": "Car_user0001_14876903_005.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "subway station", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13348 + }, + { + "path": "Kitchen_user0463_14862168_004.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "restaurant", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13349 + }, + { + "path": "Street_user0383_14827511_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "cafe", + "choice_c": "street", + "choice_d": "park", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13350 + }, + { + "path": "SubwayStation_user0487_15016470_000.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "park", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13351 + }, + { + "path": "Street_user0149_14827904_001.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "bus", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "car", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13352 + }, + { + "path": "Restaurant_user0626_15311295_005.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13353 + }, + { + "path": "Elevator_user0223_15488245_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "bus", + "choice_d": "restroom", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13354 + }, + { + "path": "Bus_user0326_14980079_001.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "street", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13355 + }, + { + "path": "Subway_user0151_14879870_001.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13356 + }, + { + "path": "Kitchen_user0087_14833019_002.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13357 + }, + { + "path": "Car_user0479_14875937_002.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13358 + }, + { + "path": "Elevator_user0356_14878228_003.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "restroom", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13359 + }, + { + "path": "ResidentialArea_user0762_14876482_001.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13360 + }, + { + "path": "CrowdedIndoor_user0411_14884535_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13361 + }, + { + "path": "Subway_user0285_14979215_001.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13362 + }, + { + "path": "Street_user0383_14827417_004.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13363 + }, + { + "path": "Subway_user0433_14869009_005.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13364 + }, + { + "path": "Street_user0761_14825328_002.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13365 + }, + { + "path": "Restroom_user0262_14829793_003.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13366 + }, + { + "path": "CrowdedIndoor_user0075_14862738_000.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13367 + }, + { + "path": "Restaurant_user0516_14987615_014.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13368 + }, + { + "path": "Elevator_user0593_15112805_001.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13369 + }, + { + "path": "Park_user0016_15053773_004.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "street", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13370 + }, + { + "path": "Park_user0173_15519005_002.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13371 + }, + { + "path": "Restaurant_user0759_14984128_006.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13372 + }, + { + "path": "ResidentialArea_user0024_14866976_005.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13373 + }, + { + "path": "Elevator_user0580_14884588_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "subway station", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13374 + }, + { + "path": "Park_user0498_15520217_001.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13375 + }, + { + "path": "Restroom_user0705_14826445_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13376 + }, + { + "path": "Cafe_user0044_15441275_005.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "restaurant", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13377 + }, + { + "path": "CrowdedIndoor_user0384_14977856_000.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13378 + }, + { + "path": "Park_user0398_15359579_000.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13379 + }, + { + "path": "Street_user0254_14823497_001.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13380 + }, + { + "path": "Restroom_user0565_14835833_005.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13381 + }, + { + "path": "Cafe_user0795_14987030_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13382 + }, + { + "path": "ResidentialArea_user0592_14866641_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13383 + }, + { + "path": "Restroom_user0262_14829793_007.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13384 + }, + { + "path": "Bus_user0762_14884557_004.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13385 + }, + { + "path": "Restroom_user0735_14837188_000.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "elevator", + "choice_b": "crowded indoor", + "choice_c": "bus", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13386 + }, + { + "path": "Subway_user0799_14979606_002.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "subway station", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13387 + }, + { + "path": "SubwayStation_user0045_15358475_005.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13388 + }, + { + "path": "Restroom_user0755_14819057_007.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "park", + "choice_b": "crowded indoor", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13389 + }, + { + "path": "Restaurant_user0755_14897368_004.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13390 + }, + { + "path": "Subway_user0092_14877179_002.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13391 + }, + { + "path": "Restroom_user0386_14849434_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13392 + }, + { + "path": "CrowdedIndoor_user0600_14986029_001.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "subway station", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13393 + }, + { + "path": "Park_user0024_15018054_004.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13394 + }, + { + "path": "ResidentialArea_user0063_14867581_001.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13395 + }, + { + "path": "Elevator_user0318_14878047_000.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13396 + }, + { + "path": "Car_user0213_14877839_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13397 + }, + { + "path": "ResidentialArea_user0536_14985908_004.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13398 + }, + { + "path": "Subway_user0231_14869776_003.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13399 + }, + { + "path": "ResidentialArea_user0352_14982346_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13400 + }, + { + "path": "Bus_user0092_14877227_006.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "subway station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13401 + }, + { + "path": "Subway_user0109_14890420_001.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13402 + }, + { + "path": "Cafe_user0460_15519441_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13403 + }, + { + "path": "Park_user0556_15520028_001.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13404 + }, + { + "path": "Park_user0024_15018064_003.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "park", + "choice_b": "subway", + "choice_c": "car", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13405 + }, + { + "path": "ResidentialArea_user0411_14866953_002.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13406 + }, + { + "path": "SubwayStation_user0132_15509322_001.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "car", + "choice_b": "subway station", + "choice_c": "bus", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13407 + }, + { + "path": "Park_user0314_15519048_004.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "subway", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13408 + }, + { + "path": "SubwayStation_user0000_15208798_005.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "park", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13409 + }, + { + "path": "ResidentialArea_user0411_14887517_002.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13410 + }, + { + "path": "Car_user0423_14861198_005.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13411 + }, + { + "path": "Cafe_user0651_15508523_000.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13412 + }, + { + "path": "SubwayStation_user0045_15119578_002.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13413 + }, + { + "path": "Park_user0807_14973602_005.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "park", + "choice_d": "restroom", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13414 + }, + { + "path": "CrowdedIndoor_user0479_14875993_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13415 + }, + { + "path": "Restroom_user0811_14846831_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13416 + }, + { + "path": "Car_user0605_14869931_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13417 + }, + { + "path": "Park_user0709_15359587_004.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13418 + }, + { + "path": "Cafe_user0089_15512146_005.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13419 + }, + { + "path": "Restroom_user0791_14816886_005.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "restroom", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13420 + }, + { + "path": "Restaurant_user0762_14884690_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13421 + }, + { + "path": "CrowdedIndoor_user0699_14877040_004.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "subway station", + "choice_b": "park", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13422 + }, + { + "path": "Restaurant_user0759_14984103_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13423 + }, + { + "path": "CrowdedIndoor_user0479_14875978_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13424 + }, + { + "path": "Elevator_user0027_15515003_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13425 + }, + { + "path": "Street_user0212_14828083_005.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "street", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13426 + }, + { + "path": "Restroom_user0811_14846831_004.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "cafe", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13427 + }, + { + "path": "Park_user0274_15053860_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13428 + }, + { + "path": "Elevator_user0277_15018035_000.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway station", + "choice_b": "elevator", + "choice_c": "cafe", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13429 + }, + { + "path": "Cafe_user0511_15113515_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "car", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13430 + }, + { + "path": "CrowdedIndoor_user0346_14888094_004.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13431 + }, + { + "path": "SubwayStation_user0685_15119584_001.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13432 + }, + { + "path": "Elevator_user0074_15396891_002.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "elevator", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13433 + }, + { + "path": "ResidentialArea_user0243_14885471_005.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13434 + }, + { + "path": "Elevator_user0040_14875611_008.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "residential area", + "choice_d": "kitchen", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13435 + }, + { + "path": "Restaurant_user0682_14882552_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13436 + }, + { + "path": "Bus_user0192_14983737_004.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "crowded indoor", + "choice_b": "car", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13437 + }, + { + "path": "Car_user0563_14875653_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13438 + }, + { + "path": "Park_user0526_15119570_003.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13439 + }, + { + "path": "Bus_user0138_14982408_001.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13440 + }, + { + "path": "Elevator_user0054_14986082_000.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "park", + "choice_b": "restaurant", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13441 + }, + { + "path": "ResidentialArea_user0133_14979289_001.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13442 + }, + { + "path": "Restaurant_user0132_14988098_002.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "subway", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13443 + }, + { + "path": "ResidentialArea_user0622_14867965_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "restaurant", + "choice_b": "crowded indoor", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13444 + }, + { + "path": "ResidentialArea_user0828_14983581_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "car", + "choice_d": "elevator", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13445 + }, + { + "path": "Restroom_user0124_14844486_000.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13446 + }, + { + "path": "CrowdedIndoor_user0500_14870561_002.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13447 + }, + { + "path": "Car_user0095_14881310_005.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13448 + }, + { + "path": "Subway_user0591_14868883_000.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13449 + }, + { + "path": "Car_user0177_14846838_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13450 + }, + { + "path": "ResidentialArea_user0670_14880716_000.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "kitchen", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13451 + }, + { + "path": "Elevator_user0110_15519028_003.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13452 + }, + { + "path": "Park_user0643_15412392_005.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13453 + }, + { + "path": "Restroom_user0417_14843636_004.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "cafe", + "choice_d": "car", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13454 + }, + { + "path": "Car_user0049_14880117_004.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway station", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13455 + }, + { + "path": "Bus_user0138_14982408_003.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "park", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13456 + }, + { + "path": "Park_user0709_15359575_000.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13457 + }, + { + "path": "Restroom_user0248_14847790_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "restroom", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13458 + }, + { + "path": "Kitchen_user0592_14860575_003.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13459 + }, + { + "path": "Kitchen_user0066_14862259_004.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13460 + }, + { + "path": "Bus_user0344_14873381_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13461 + }, + { + "path": "Car_user0007_14855582_001.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13462 + }, + { + "path": "Kitchen_user0585_14834777_003.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13463 + }, + { + "path": "Car_user0592_14862162_000.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13464 + }, + { + "path": "Restroom_user0464_14835121_001.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13465 + }, + { + "path": "Subway_user0435_14876255_004.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13466 + }, + { + "path": "Bus_user0272_14882829_001.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "kitchen", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13467 + }, + { + "path": "Restroom_user0262_14829793_012.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13468 + }, + { + "path": "Car_user0201_14883783_006.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13469 + }, + { + "path": "Car_user0682_14875419_004.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13470 + }, + { + "path": "Elevator_user0822_15119162_004.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "restroom", + "choice_d": "residential area", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13471 + }, + { + "path": "SubwayStation_user0482_15022408_002.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "subway station", + "choice_d": "park", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13472 + }, + { + "path": "Street_user0712_14832917_001.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13473 + }, + { + "path": "Park_user0460_15519894_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13474 + }, + { + "path": "Restaurant_user0226_14982495_004.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13475 + }, + { + "path": "Cafe_user0650_15304616_004.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13476 + }, + { + "path": "Elevator_user0526_15119351_003.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13477 + }, + { + "path": "ResidentialArea_user0411_14867588_005.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "restaurant", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13478 + }, + { + "path": "CrowdedIndoor_user0522_14980209_005.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "elevator", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13479 + }, + { + "path": "SubwayStation_user0132_15508860_004.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13480 + }, + { + "path": "ResidentialArea_user0817_14984040_004.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13481 + }, + { + "path": "Subway_user0435_14876229_005.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13482 + }, + { + "path": "ResidentialArea_user0308_14868944_002.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13483 + }, + { + "path": "Elevator_user0386_14870200_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13484 + }, + { + "path": "ResidentialArea_user0386_14863004_005.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "street", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13485 + }, + { + "path": "Car_user0692_14853357_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13486 + }, + { + "path": "Bus_user0009_14981454_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "kitchen", + "choice_d": "subway", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13487 + }, + { + "path": "Kitchen_user0277_14860134_003.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "subway", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13488 + }, + { + "path": "Cafe_user0074_15519053_005.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13489 + }, + { + "path": "Car_user0656_14852905_003.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "car", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13490 + }, + { + "path": "Elevator_user0110_15519028_000.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "restroom", + "choice_b": "cafe", + "choice_c": "restaurant", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13491 + }, + { + "path": "Elevator_user0698_15505461_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13492 + }, + { + "path": "Bus_user0138_14982391_017.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13493 + }, + { + "path": "Cafe_user0265_15519064_002.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "crowded indoor", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13494 + }, + { + "path": "Bus_user0272_14882834_005.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "bus", + "choice_d": "subway station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13495 + }, + { + "path": "Bus_user0017_14896159_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "bus", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13496 + }, + { + "path": "Restroom_user0659_14823293_002.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13497 + }, + { + "path": "Subway_user0118_14886491_004.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "car", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13498 + }, + { + "path": "Restaurant_user0498_14880359_001.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "restaurant", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13499 + }, + { + "path": "Street_user0530_14835161_003.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "cafe", + "choice_d": "subway station", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13500 + }, + { + "path": "Restroom_user0365_14830208_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13501 + }, + { + "path": "Car_user0180_14861260_002.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13502 + }, + { + "path": "ResidentialArea_user0828_14983581_001.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "subway", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13503 + }, + { + "path": "CrowdedIndoor_user0235_14876816_000.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13504 + }, + { + "path": "Restaurant_user0759_14984131_001.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "subway station", + "choice_b": "cafe", + "choice_c": "restaurant", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13505 + }, + { + "path": "Kitchen_user0601_14852290_007.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "restaurant", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13506 + }, + { + "path": "Restaurant_user0516_14987615_004.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13507 + }, + { + "path": "Park_user0429_14977599_004.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13508 + }, + { + "path": "Restaurant_user0759_14984129_000.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "park", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "kitchen", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13509 + }, + { + "path": "Bus_user0330_14875707_003.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13510 + }, + { + "path": "Bus_user0658_14869138_005.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13511 + }, + { + "path": "Subway_user0433_14875693_003.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe", + "choice_b": "subway", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13512 + }, + { + "path": "Kitchen_user0688_14838296_004.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "crowded indoor", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13513 + }, + { + "path": "Kitchen_user0066_14862259_002.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13514 + }, + { + "path": "Street_user0183_14830440_002.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13515 + }, + { + "path": "Street_user0383_14828691_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13516 + }, + { + "path": "SubwayStation_user0395_15119410_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13517 + }, + { + "path": "Subway_user0151_14879892_005.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13518 + }, + { + "path": "SubwayStation_user0378_14986010_005.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13519 + }, + { + "path": "Park_user0698_15053571_003.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13520 + }, + { + "path": "Elevator_user0527_14871672_004.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13521 + }, + { + "path": "ResidentialArea_user0081_14895415_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "subway station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13522 + }, + { + "path": "CrowdedIndoor_user0015_14870152_001.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13523 + }, + { + "path": "CrowdedIndoor_user0346_14888131_001.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway station", + "choice_b": "crowded indoor", + "choice_c": "restaurant", + "choice_d": "residential area", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13524 + }, + { + "path": "CrowdedIndoor_user0461_14894927_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "crowded indoor", + "choice_b": "car", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13525 + }, + { + "path": "Bus_user0530_14878241_000.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13526 + }, + { + "path": "SubwayStation_user0291_14885689_002.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "subway", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13527 + }, + { + "path": "Restroom_user0235_14817123_000.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "street", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13528 + }, + { + "path": "Elevator_user0386_14870205_000.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13529 + }, + { + "path": "SubwayStation_user0290_15519808_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "elevator", + "choice_d": "residential area", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13530 + }, + { + "path": "Restroom_user0368_14852233_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13531 + }, + { + "path": "Restroom_user0473_14817076_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13532 + }, + { + "path": "Kitchen_user0589_14842032_014.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13533 + }, + { + "path": "Elevator_user0317_14987346_004.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "restroom", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13534 + }, + { + "path": "Bus_user0009_14982091_005.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "elevator", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13535 + }, + { + "path": "Street_user0462_14827358_004.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13536 + }, + { + "path": "SubwayStation_user0416_15519120_005.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13537 + }, + { + "path": "ResidentialArea_user0167_14984233_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13538 + }, + { + "path": "Street_user0712_14832996_004.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13539 + }, + { + "path": "SubwayStation_user0277_14880137_001.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13540 + }, + { + "path": "Cafe_user0077_15441245_000.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13541 + }, + { + "path": "SubwayStation_user0357_15119165_002.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13542 + }, + { + "path": "Restaurant_user0799_15039676_005.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13543 + }, + { + "path": "Park_user0334_15519083_003.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "crowded indoor", + "choice_d": "kitchen", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13544 + }, + { + "path": "Subway_user0433_14875693_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13545 + }, + { + "path": "Kitchen_user0219_14827248_003.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13546 + }, + { + "path": "Elevator_user0580_14884588_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "cafe", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13547 + }, + { + "path": "Car_user0453_14882579_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13548 + }, + { + "path": "Kitchen_user0205_14818201_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "crowded indoor", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13549 + }, + { + "path": "SubwayStation_user0000_15208798_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "street", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "restaurant", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13550 + }, + { + "path": "Restaurant_user0132_14988094_000.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13551 + }, + { + "path": "Restaurant_user0151_14879938_003.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13552 + }, + { + "path": "Park_user0314_14893811_003.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13553 + }, + { + "path": "Park_user0596_15440296_005.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "cafe", + "choice_b": "elevator", + "choice_c": "park", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13554 + }, + { + "path": "Cafe_user0802_14892056_004.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13555 + }, + { + "path": "ResidentialArea_user0684_14977450_006.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13556 + }, + { + "path": "Street_user0789_14825445_003.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "subway station", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13557 + }, + { + "path": "Bus_user0408_14867711_003.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13558 + }, + { + "path": "Park_user0274_15053856_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13559 + }, + { + "path": "Kitchen_user0599_14822265_002.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13560 + }, + { + "path": "ResidentialArea_user0811_14979386_001.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13561 + }, + { + "path": "Subway_user0507_14869096_005.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13562 + }, + { + "path": "Kitchen_user0445_14861774_004.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "restroom", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13563 + }, + { + "path": "Cafe_user0433_15311928_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "subway", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13564 + }, + { + "path": "Restaurant_user0755_14879459_008.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13565 + }, + { + "path": "SubwayStation_user0438_15119086_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13566 + }, + { + "path": "Car_user0177_14846838_000.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13567 + }, + { + "path": "Street_user0408_14826554_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13568 + }, + { + "path": "ResidentialArea_user0628_14982958_004.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13569 + }, + { + "path": "Restroom_user0789_14825848_003.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13570 + }, + { + "path": "Cafe_user0105_14882799_006.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "subway", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13571 + }, + { + "path": "Cafe_user0598_15518991_003.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13572 + }, + { + "path": "Restaurant_user0567_14894339_004.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13573 + }, + { + "path": "Elevator_user0239_14985952_002.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "elevator", + "choice_d": "cafe", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13574 + }, + { + "path": "Park_user0238_15519568_004.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13575 + }, + { + "path": "Car_user0753_14867137_007.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "restaurant", + "choice_b": "subway", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13576 + }, + { + "path": "Car_user0070_14855308_001.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13577 + }, + { + "path": "SubwayStation_user0416_15519120_000.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13578 + }, + { + "path": "Bus_user0618_14895179_004.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13579 + }, + { + "path": "Bus_user0762_14885410_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13580 + }, + { + "path": "Elevator_user0397_15134570_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "elevator", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13581 + }, + { + "path": "Bus_user0330_14875732_000.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "elevator", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13582 + }, + { + "path": "Bus_user0151_14879810_004.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13583 + }, + { + "path": "CrowdedIndoor_user0585_14899066_004.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13584 + }, + { + "path": "Cafe_user0582_15053531_003.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "restaurant", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "car", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13585 + }, + { + "path": "Restroom_user0245_14831582_000.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13586 + }, + { + "path": "Bus_user0530_14878239_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "park", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13587 + }, + { + "path": "Cafe_user0266_14977244_005.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13588 + }, + { + "path": "Cafe_user0027_15311842_002.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "park", + "choice_d": "crowded indoor", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13589 + }, + { + "path": "ResidentialArea_user0014_14983373_002.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13590 + }, + { + "path": "ResidentialArea_user0670_14880710_002.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13591 + }, + { + "path": "Subway_user0285_14979215_002.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "subway station", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13592 + }, + { + "path": "Restaurant_user0571_15519150_005.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13593 + }, + { + "path": "SubwayStation_user0403_15486004_003.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "cafe", + "choice_b": "subway station", + "choice_c": "restaurant", + "choice_d": "restroom", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13594 + }, + { + "path": "Restroom_user0235_14816993_004.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "restroom", + "choice_b": "elevator", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13595 + }, + { + "path": "Street_user0383_14828754_002.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "street", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13596 + }, + { + "path": "Elevator_user0318_14878047_005.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "cafe", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13597 + }, + { + "path": "Car_user0494_14853999_002.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "residential area", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13598 + }, + { + "path": "Elevator_user0680_14983889_005.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13599 + }, + { + "path": "Restroom_user0036_14830609_002.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13600 + }, + { + "path": "Car_user0056_14874961_005.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "street", + "choice_b": "crowded indoor", + "choice_c": "restroom", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13601 + }, + { + "path": "ResidentialArea_user0670_14880716_005.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13602 + }, + { + "path": "Kitchen_user0445_14861699_004.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "elevator", + "choice_b": "kitchen", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13603 + }, + { + "path": "Subway_user0116_14895124_001.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13604 + }, + { + "path": "CrowdedIndoor_user0153_14980692_002.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13605 + }, + { + "path": "CrowdedIndoor_user0707_14868496_004.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "bus", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13606 + }, + { + "path": "Park_user0238_15519568_005.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "park", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13607 + }, + { + "path": "Kitchen_user0107_14853120_005.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "subway station", + "choice_b": "elevator", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13608 + }, + { + "path": "Park_user0808_15507861_003.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "subway station", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13609 + }, + { + "path": "CrowdedIndoor_user0479_14875993_003.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "subway", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13610 + }, + { + "path": "Restaurant_user0516_14987617_002.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "restaurant", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "subway", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13611 + }, + { + "path": "Subway_user0820_14892867_001.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13612 + }, + { + "path": "Restaurant_user0799_14981029_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "restroom", + "choice_b": "residential area", + "choice_c": "restaurant", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13613 + }, + { + "path": "Kitchen_user0585_14834760_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "cafe", + "choice_b": "kitchen", + "choice_c": "crowded indoor", + "choice_d": "subway station", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13614 + }, + { + "path": "Subway_user0152_14869815_001.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "bus", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13615 + }, + { + "path": "SubwayStation_user0017_15311279_000.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "elevator", + "choice_b": "restaurant", + "choice_c": "cafe", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13616 + }, + { + "path": "Restaurant_user0045_15053600_003.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13617 + }, + { + "path": "Restroom_user0140_14830538_003.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "street", + "choice_b": "subway", + "choice_c": "restroom", + "choice_d": "residential area", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13618 + }, + { + "path": "Kitchen_user0016_14871485_005.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "crowded indoor", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13619 + }, + { + "path": "SubwayStation_user0389_15519832_001.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "cafe", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13620 + }, + { + "path": "ResidentialArea_user0024_14868790_002.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "restroom", + "choice_b": "cafe", + "choice_c": "restaurant", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13621 + }, + { + "path": "Bus_user0009_14897252_005.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13622 + }, + { + "path": "Restaurant_user0799_15039676_000.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "restaurant", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13623 + }, + { + "path": "Cafe_user0650_15518735_000.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13624 + }, + { + "path": "Kitchen_user0370_14853988_003.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13625 + }, + { + "path": "Street_user0524_14835809_001.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13626 + }, + { + "path": "CrowdedIndoor_user0699_14877008_005.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13627 + }, + { + "path": "Elevator_user0503_14979170_002.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway station", + "choice_b": "restaurant", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13628 + }, + { + "path": "Cafe_user0544_14985158_005.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "subway station", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13629 + }, + { + "path": "ResidentialArea_user0235_14868431_000.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13630 + }, + { + "path": "Park_user0498_15119288_004.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13631 + }, + { + "path": "Car_user0453_14882539_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "street", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13632 + }, + { + "path": "CrowdedIndoor_user0707_14868508_002.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13633 + }, + { + "path": "SubwayStation_user0044_15519156_004.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "bus", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13634 + }, + { + "path": "Street_user0149_14827904_003.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "cafe", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "bus", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13635 + }, + { + "path": "Subway_user0293_14880859_001.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13636 + }, + { + "path": "Kitchen_user0365_14830066_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13637 + }, + { + "path": "Car_user0729_14859353_005.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "kitchen", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13638 + }, + { + "path": "Restroom_user0417_14843636_001.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13639 + }, + { + "path": "Restaurant_user0423_14877986_000.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13640 + }, + { + "path": "Car_user0605_14869917_005.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13641 + }, + { + "path": "Park_user0208_15519476_002.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13642 + }, + { + "path": "Subway_user0791_14879526_003.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "car", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13643 + }, + { + "path": "Elevator_user0239_14986841_003.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13644 + }, + { + "path": "Kitchen_user0595_14838929_004.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "residential area", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13645 + }, + { + "path": "Elevator_user0680_14983889_000.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13646 + }, + { + "path": "Bus_user0704_14884875_000.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13647 + }, + { + "path": "Elevator_user0210_15400852_000.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "subway station", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13648 + }, + { + "path": "Car_user0047_14861777_005.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13649 + }, + { + "path": "Street_user0789_14841195_001.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13650 + }, + { + "path": "Cafe_user0266_15293781_002.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "street", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13651 + }, + { + "path": "Street_user0789_14841151_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13652 + }, + { + "path": "Subway_user0148_14897664_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13653 + }, + { + "path": "Elevator_user0132_15508932_004.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13654 + }, + { + "path": "Street_user0108_14840677_003.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13655 + }, + { + "path": "Park_user0016_15295247_003.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13656 + }, + { + "path": "Cafe_user0620_14894707_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "restroom", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "crowded indoor", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13657 + }, + { + "path": "Car_user0453_14882571_000.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13658 + }, + { + "path": "Park_user0274_15053854_001.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "park", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "subway station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13659 + }, + { + "path": "Car_user0220_14876413_004.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13660 + }, + { + "path": "SubwayStation_user0277_14880125_003.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13661 + }, + { + "path": "Elevator_user0596_15119384_005.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "cafe", + "choice_b": "car", + "choice_c": "elevator", + "choice_d": "subway station", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13662 + }, + { + "path": "Park_user0741_15518846_004.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13663 + }, + { + "path": "Cafe_user0609_14888282_000.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "kitchen", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13664 + }, + { + "path": "Subway_user0635_14981759_004.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "kitchen", + "choice_b": "crowded indoor", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13665 + }, + { + "path": "Street_user0817_14840622_005.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13666 + }, + { + "path": "Restroom_user0012_14818556_013.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13667 + }, + { + "path": "ResidentialArea_user0081_14875972_001.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13668 + }, + { + "path": "Restaurant_user0132_14988093_004.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "restaurant", + "choice_b": "elevator", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13669 + }, + { + "path": "Street_user0235_14835163_003.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13670 + }, + { + "path": "Cafe_user0620_14894702_000.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "park", + "choice_b": "crowded indoor", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13671 + }, + { + "path": "Elevator_user0709_15359569_002.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13672 + }, + { + "path": "Cafe_user0808_15519080_005.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13673 + }, + { + "path": "ResidentialArea_user0503_15016495_003.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13674 + }, + { + "path": "Bus_user0272_14882820_003.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "restaurant", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13675 + }, + { + "path": "CrowdedIndoor_user0585_14899833_000.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "subway station", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13676 + }, + { + "path": "ResidentialArea_user0165_14862164_005.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13677 + }, + { + "path": "Park_user0024_15018062_003.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "elevator", + "choice_b": "restaurant", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13678 + }, + { + "path": "Restroom_user0453_14831184_004.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "kitchen", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13679 + }, + { + "path": "Cafe_user0626_15311290_001.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "car", + "choice_b": "restaurant", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13680 + }, + { + "path": "Park_user0821_15359462_001.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "subway", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13681 + }, + { + "path": "Car_user0001_14876917_001.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13682 + }, + { + "path": "Car_user0494_14853999_001.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "kitchen", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "restaurant", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13683 + }, + { + "path": "CrowdedIndoor_user0699_14877032_002.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "cafe", + "choice_b": "restroom", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13684 + }, + { + "path": "Restroom_user0463_14844305_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13685 + }, + { + "path": "Cafe_user0597_14884544_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "restaurant", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13686 + }, + { + "path": "Cafe_user0444_15053221_004.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "cafe", + "choice_b": "park", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13687 + }, + { + "path": "ResidentialArea_user0777_14983232_003.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "restroom", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13688 + }, + { + "path": "ResidentialArea_user0684_14977450_002.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "park", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13689 + }, + { + "path": "Restroom_user0368_14852233_001.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "restroom", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13690 + }, + { + "path": "ResidentialArea_user0680_14983897_000.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "crowded indoor", + "choice_d": "restroom", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13691 + }, + { + "path": "Car_user0729_14858760_003.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "elevator", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13692 + }, + { + "path": "CrowdedIndoor_user0356_14867614_001.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "subway station", + "choice_b": "crowded indoor", + "choice_c": "street", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13693 + }, + { + "path": "Restaurant_user0089_15053646_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13694 + }, + { + "path": "SubwayStation_user0138_14984630_009.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13695 + }, + { + "path": "Elevator_user0210_15400852_005.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "subway station", + "choice_c": "elevator", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13696 + }, + { + "path": "Cafe_user0416_15519127_000.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13697 + }, + { + "path": "Car_user0563_14875677_004.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13698 + }, + { + "path": "Car_user0369_14868282_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13699 + }, + { + "path": "Elevator_user0476_14986657_002.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "street", + "choice_d": "car", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13700 + }, + { + "path": "Kitchen_user0316_14850331_000.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "cafe", + "choice_b": "subway station", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13701 + }, + { + "path": "Street_user0314_14822536_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "restroom", + "choice_b": "street", + "choice_c": "kitchen", + "choice_d": "cafe", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13702 + }, + { + "path": "Bus_user0820_14885789_000.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13703 + }, + { + "path": "Bus_user0151_14879799_004.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "elevator", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13704 + }, + { + "path": "CrowdedIndoor_user0378_14986012_003.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "subway station", + "choice_b": "subway", + "choice_c": "residential area", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13705 + }, + { + "path": "Car_user0159_14867880_002.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13706 + }, + { + "path": "Kitchen_user0286_14857125_000.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "kitchen", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "street", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13707 + }, + { + "path": "ResidentialArea_user0308_14868725_005.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "residential area", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13708 + }, + { + "path": "Street_user0235_14834803_000.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "residential area", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13709 + }, + { + "path": "Car_user0129_14869765_002.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "residential area", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13710 + }, + { + "path": "Cafe_user0001_14981568_005.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13711 + }, + { + "path": "Car_user0731_14867196_002.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13712 + }, + { + "path": "Kitchen_user0760_14833035_004.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "elevator", + "choice_b": "restaurant", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13713 + }, + { + "path": "Subway_user0285_14979210_003.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "restaurant", + "choice_b": "street", + "choice_c": "park", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13714 + }, + { + "path": "Cafe_user0596_15311214_002.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "subway station", + "choice_b": "park", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13715 + }, + { + "path": "Bus_user0200_14875469_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "bus", + "choice_b": "restaurant", + "choice_c": "elevator", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13716 + }, + { + "path": "Cafe_user0718_15518852_001.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "subway station", + "choice_d": "kitchen", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13717 + }, + { + "path": "SubwayStation_user0045_15311828_004.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "car", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13718 + }, + { + "path": "Elevator_user0099_15113477_002.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "residential area", + "choice_b": "elevator", + "choice_c": "street", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13719 + }, + { + "path": "Elevator_user0349_14982235_000.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13720 + }, + { + "path": "Street_user0712_14821785_002.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "subway", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "restroom", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13721 + }, + { + "path": "CrowdedIndoor_user0346_14888131_004.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "residential area", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13722 + }, + { + "path": "Subway_user0557_14872046_003.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "restroom", + "choice_b": "subway", + "choice_c": "kitchen", + "choice_d": "restaurant", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13723 + }, + { + "path": "Street_user0181_14846735_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "street", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13724 + }, + { + "path": "Restaurant_user0151_14879944_005.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "restaurant", + "choice_d": "restroom", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13725 + }, + { + "path": "Car_user0177_14853931_000.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13726 + }, + { + "path": "Street_user0575_14833204_001.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "park", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "residential area", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13727 + }, + { + "path": "CrowdedIndoor_user0500_14870561_004.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13728 + }, + { + "path": "Bus_user0326_14870508_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway station", + "choice_b": "elevator", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13729 + }, + { + "path": "Cafe_user0718_15518858_004.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "street", + "choice_d": "subway", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13730 + }, + { + "path": "Restroom_user0369_14838373_003.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "elevator", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13731 + }, + { + "path": "Car_user0729_14862037_004.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13732 + }, + { + "path": "SubwayStation_user0799_15011492_003.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "subway station", + "choice_d": "elevator", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13733 + }, + { + "path": "Restroom_user0188_14836530_000.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "restroom", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13734 + }, + { + "path": "Car_user0001_14876921_000.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13735 + }, + { + "path": "Restroom_user0013_14833351_001.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "residential area", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13736 + }, + { + "path": "Park_user0709_15359585_002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13737 + }, + { + "path": "Restroom_user0735_14845589_004.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13738 + }, + { + "path": "Subway_user0092_14877130_003.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "elevator", + "choice_d": "residential area", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13739 + }, + { + "path": "Street_user0346_14824951_004.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "car", + "choice_b": "crowded indoor", + "choice_c": "street", + "choice_d": "elevator", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13740 + }, + { + "path": "ResidentialArea_user0536_14985895_002.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13741 + }, + { + "path": "Elevator_user0593_14878960_004.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "crowded indoor", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13742 + }, + { + "path": "Subway_user0791_14879547_000.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "restroom", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13743 + }, + { + "path": "Restaurant_user0652_14981631_004.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "kitchen", + "choice_d": "subway station", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13744 + }, + { + "path": "Car_user0563_14875667_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13745 + }, + { + "path": "Kitchen_user0066_14862275_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13746 + }, + { + "path": "ResidentialArea_user0536_14985895_000.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "crowded indoor", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13747 + }, + { + "path": "SubwayStation_user0416_15519120_001.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "crowded indoor", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13748 + }, + { + "path": "Restroom_user0188_14836530_006.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "subway", + "choice_d": "subway station", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13749 + }, + { + "path": "ResidentialArea_user0243_14885471_002.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "cafe", + "choice_d": "kitchen", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13750 + }, + { + "path": "Car_user0777_14867611_001.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13751 + }, + { + "path": "Bus_user0530_14878216_001.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13752 + }, + { + "path": "Kitchen_user0646_14833246_001.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "residential area", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13753 + }, + { + "path": "Park_user0308_14978497_005.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "restaurant", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13754 + }, + { + "path": "Car_user0605_14869939_004.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13755 + }, + { + "path": "Park_user0698_15518993_001.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "cafe", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13756 + }, + { + "path": "Elevator_user0651_15053912_004.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "subway", + "choice_b": "cafe", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13757 + }, + { + "path": "CrowdedIndoor_user0789_14869235_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "elevator", + "choice_b": "subway station", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13758 + }, + { + "path": "Bus_user0152_14880041_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "bus", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13759 + }, + { + "path": "Restaurant_user0533_15119121_002.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13760 + }, + { + "path": "Street_user0383_14828643_005.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13761 + }, + { + "path": "Subway_user0435_14876238_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "street", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13762 + }, + { + "path": "Elevator_user0593_15112805_005.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "bus", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13763 + }, + { + "path": "SubwayStation_user0132_15508875_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "car", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13764 + }, + { + "path": "Subway_user0435_14876220_005.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "park", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13765 + }, + { + "path": "Elevator_user0460_15519454_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "crowded indoor", + "choice_b": "residential area", + "choice_c": "cafe", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13766 + }, + { + "path": "Elevator_user0438_15295331_000.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "elevator", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "restaurant", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13767 + }, + { + "path": "Restaurant_user0334_15519025_002.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13768 + }, + { + "path": "Subway_user0395_14867643_005.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "residential area", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13769 + }, + { + "path": "Subway_user0812_14890365_000.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "residential area", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13770 + }, + { + "path": "Park_user0024_15018058_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "crowded indoor", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13771 + }, + { + "path": "Kitchen_user0585_14834791_001.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "bus", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13772 + }, + { + "path": "ResidentialArea_user0684_14977450_004.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "street", + "choice_b": "subway station", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13773 + }, + { + "path": "Bus_user0380_14869637_004.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "subway station", + "choice_c": "cafe", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13774 + }, + { + "path": "Cafe_user0057_14977311_004.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "kitchen", + "choice_b": "subway", + "choice_c": "subway station", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13775 + }, + { + "path": "Cafe_user0057_14980614_005.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13776 + }, + { + "path": "Restaurant_user0334_15519023_005.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "bus", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13777 + }, + { + "path": "Restaurant_user0044_15441273_001.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "subway", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13778 + }, + { + "path": "Subway_user0749_14981956_003.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "park", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13779 + }, + { + "path": "Park_user0596_15440295_001.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "kitchen", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13780 + }, + { + "path": "Subway_user0109_14890467_002.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "bus", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13781 + }, + { + "path": "Restaurant_user0759_14984122_001.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway station", + "choice_b": "crowded indoor", + "choice_c": "cafe", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13782 + }, + { + "path": "Kitchen_user0808_14834648_002.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "kitchen", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13783 + }, + { + "path": "Elevator_user0165_14875749_000.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "residential area", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13784 + }, + { + "path": "Bus_user0151_14879802_005.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13785 + }, + { + "path": "Kitchen_user0653_14861242_001.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13786 + }, + { + "path": "Elevator_user0657_14873264_002.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restaurant", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13787 + }, + { + "path": "Bus_user0704_14884958_002.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13788 + }, + { + "path": "Elevator_user0433_14883061_000.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway", + "choice_b": "elevator", + "choice_c": "kitchen", + "choice_d": "restroom", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13789 + }, + { + "path": "Car_user0753_14867137_002.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13790 + }, + { + "path": "Elevator_user0651_15053969_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "elevator", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "restroom", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13791 + }, + { + "path": "Car_user0453_14882554_000.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "restroom", + "choice_b": "kitchen", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13792 + }, + { + "path": "Subway_user0148_14897618_002.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "crowded indoor", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13793 + }, + { + "path": "Street_user0817_14840616_005.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "street", + "choice_d": "car", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13794 + }, + { + "path": "ResidentialArea_user0167_14984233_000.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "subway", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "street", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13795 + }, + { + "path": "Car_user0213_14877842_005.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "restroom", + "choice_b": "bus", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13796 + }, + { + "path": "Car_user0729_14858787_004.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "elevator", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13797 + }, + { + "path": "Subway_user0293_14894190_003.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "bus", + "choice_b": "elevator", + "choice_c": "subway", + "choice_d": "crowded indoor", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13798 + }, + { + "path": "Subway_user0528_14980298_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "subway", + "choice_b": "park", + "choice_c": "cafe", + "choice_d": "restroom", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13799 + }, + { + "path": "SubwayStation_user0709_15508733_001.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "elevator", + "choice_b": "crowded indoor", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13800 + }, + { + "path": "Restaurant_user0500_15518830_005.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "kitchen", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13801 + }, + { + "path": "CrowdedIndoor_user0585_14879790_005.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "kitchen", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "car", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13802 + }, + { + "path": "Elevator_user0822_15119162_001.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "street", + "choice_b": "bus", + "choice_c": "elevator", + "choice_d": "kitchen", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13803 + }, + { + "path": "Subway_user0293_14894049_000.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "park", + "choice_b": "cafe", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13804 + }, + { + "path": "Subway_user0435_14876255_001.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "kitchen", + "choice_c": "cafe", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13805 + }, + { + "path": "Elevator_user0596_15119386_000.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "subway station", + "choice_d": "subway", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13806 + }, + { + "path": "Street_user0235_14827881_002.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "restaurant", + "choice_d": "car", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13807 + }, + { + "path": "Subway_user0318_14875379_003.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13808 + }, + { + "path": "ResidentialArea_user0167_14984242_000.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "residential area", + "choice_b": "subway station", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13809 + }, + { + "path": "Park_user0308_14980341_003.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "subway station", + "choice_b": "park", + "choice_c": "street", + "choice_d": "crowded indoor", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13810 + }, + { + "path": "Restaurant_user0334_15519022_000.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "elevator", + "choice_b": "car", + "choice_c": "crowded indoor", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13811 + }, + { + "path": "Subway_user0591_14868917_004.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe", + "choice_b": "subway station", + "choice_c": "subway", + "choice_d": "elevator", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13812 + }, + { + "path": "SubwayStation_user0528_14980286_002.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "kitchen", + "choice_b": "subway station", + "choice_c": "restroom", + "choice_d": "restaurant", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13813 + }, + { + "path": "Bus_user0293_14879014_003.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "restaurant", + "choice_b": "bus", + "choice_c": "subway station", + "choice_d": "street", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13814 + }, + { + "path": "Bus_user0762_14884557_001.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "subway", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "cafe", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13815 + }, + { + "path": "Kitchen_user0404_14864421_003.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "restaurant", + "choice_b": "kitchen", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13816 + }, + { + "path": "Park_user0024_15018065_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "cafe", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13817 + }, + { + "path": "ResidentialArea_user0014_14983392_005.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "residential area", + "choice_b": "restaurant", + "choice_c": "crowded indoor", + "choice_d": "cafe", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13818 + }, + { + "path": "Street_user0426_14844411_000.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13819 + }, + { + "path": "Restroom_user0188_14836498_004.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "street", + "choice_b": "residential area", + "choice_c": "subway", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13820 + }, + { + "path": "Restroom_user0585_14835132_000.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "subway", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13821 + }, + { + "path": "Bus_user0618_14892133_005.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "kitchen", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13822 + }, + { + "path": "Bus_user0272_14882717_000.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "elevator", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13823 + }, + { + "path": "Restroom_user0124_14844517_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "crowded indoor", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "kitchen", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13824 + }, + { + "path": "Subway_user0491_14869748_003.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "subway", + "choice_b": "bus", + "choice_c": "street", + "choice_d": "restroom", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13825 + }, + { + "path": "CrowdedIndoor_user0479_14875983_002.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "restroom", + "choice_b": "crowded indoor", + "choice_c": "restaurant", + "choice_d": "kitchen", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13826 + }, + { + "path": "Restaurant_user0099_15122216_003.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "restaurant", + "choice_b": "car", + "choice_c": "park", + "choice_d": "elevator", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13827 + }, + { + "path": "Kitchen_user0760_14833035_002.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "cafe", + "choice_b": "kitchen", + "choice_c": "subway station", + "choice_d": "restroom", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13828 + }, + { + "path": "CrowdedIndoor_user0712_14867431_004.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "crowded indoor", + "choice_d": "park", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13829 + }, + { + "path": "SubwayStation_user0482_15022794_004.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "restroom", + "choice_b": "restaurant", + "choice_c": "subway station", + "choice_d": "car", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13830 + }, + { + "path": "Subway_user0426_14873037_003.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "subway station", + "choice_b": "park", + "choice_c": "subway", + "choice_d": "kitchen", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13831 + }, + { + "path": "Kitchen_user0386_14817863_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "crowded indoor", + "choice_b": "cafe", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13832 + }, + { + "path": "Park_user0076_14987381_005.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "restroom", + "choice_d": "street", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13833 + }, + { + "path": "Kitchen_user0658_14836362_001.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "restaurant", + "choice_b": "subway station", + "choice_c": "street", + "choice_d": "kitchen", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13834 + }, + { + "path": "Kitchen_user0772_14855822_001.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "subway", + "choice_b": "restaurant", + "choice_c": "kitchen", + "choice_d": "bus", + "answer_gt": "kitchen", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13835 + }, + { + "path": "Cafe_user0620_14894696_003.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13836 + }, + { + "path": "Car_user0729_14858760_000.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13837 + }, + { + "path": "Cafe_user0074_15396574_005.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "restaurant", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "elevator", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13838 + }, + { + "path": "Restroom_user0089_14816246_001.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "restaurant", + "choice_b": "street", + "choice_c": "car", + "choice_d": "restroom", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13839 + }, + { + "path": "Restaurant_user0762_14884670_000.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "street", + "choice_c": "restroom", + "choice_d": "restaurant", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13840 + }, + { + "path": "SubwayStation_user0482_14997176_004.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "kitchen", + "choice_b": "bus", + "choice_c": "restroom", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13841 + }, + { + "path": "Park_user0526_15119570_004.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "restaurant", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "crowded indoor", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13842 + }, + { + "path": "Elevator_user0004_15441125_003.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "street", + "choice_b": "elevator", + "choice_c": "restaurant", + "choice_d": "subway station", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13843 + }, + { + "path": "CrowdedIndoor_user0188_14863171_003.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "crowded indoor", + "choice_b": "kitchen", + "choice_c": "elevator", + "choice_d": "restaurant", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13844 + }, + { + "path": "ResidentialArea_user0339_14871472_005.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13845 + }, + { + "path": "Restroom_user0614_14817960_003.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "park", + "choice_b": "elevator", + "choice_c": "restroom", + "choice_d": "subway", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13846 + }, + { + "path": "CrowdedIndoor_user0116_14869786_002.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "cafe", + "choice_b": "crowded indoor", + "choice_c": "elevator", + "choice_d": "street", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13847 + }, + { + "path": "Subway_user0635_14981549_001.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "elevator", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13848 + }, + { + "path": "Elevator_user0305_15311072_000.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "kitchen", + "choice_d": "elevator", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13849 + }, + { + "path": "Street_user0408_14826521_000.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe", + "choice_b": "street", + "choice_c": "residential area", + "choice_d": "subway", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13850 + }, + { + "path": "Park_user0067_15519727_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "restroom", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13851 + }, + { + "path": "SubwayStation_user0528_14980290_002.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "subway station", + "choice_b": "kitchen", + "choice_c": "residential area", + "choice_d": "street", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13852 + }, + { + "path": "ResidentialArea_user0643_15018022_005.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "car", + "choice_b": "cafe", + "choice_c": "residential area", + "choice_d": "restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13853 + }, + { + "path": "Street_user0417_14844681_004.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "restroom", + "choice_c": "street", + "choice_d": "subway station", + "answer_gt": "street", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13854 + }, + { + "path": "Cafe_user0017_15310943_001.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "cafe", + "choice_d": "park", + "answer_gt": "cafe", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13855 + }, + { + "path": "Car_user0369_14868282_000.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "crowded indoor", + "choice_b": "car", + "choice_c": "restroom", + "choice_d": "kitchen", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13856 + }, + { + "path": "Bus_user0009_14897255_001.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "car", + "choice_b": "subway", + "choice_c": "bus", + "choice_d": "cafe", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13857 + }, + { + "path": "ResidentialArea_user0378_14869326_005.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "subway", + "choice_b": "crowded indoor", + "choice_c": "restaurant", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13858 + }, + { + "path": "Restaurant_user0684_14983584_005.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13859 + }, + { + "path": "Subway_user0109_14890449_003.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "elevator", + "choice_b": "subway", + "choice_c": "restroom", + "choice_d": "cafe", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13860 + }, + { + "path": "Bus_user0200_14875499_001.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "subway station", + "choice_b": "street", + "choice_c": "bus", + "choice_d": "crowded indoor", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13861 + }, + { + "path": "Subway_user0065_14870466_000.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "elevator", + "choice_d": "subway", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13862 + }, + { + "path": "ResidentialArea_user0148_14970247_004.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "street", + "choice_b": "cafe", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13863 + }, + { + "path": "CrowdedIndoor_user0216_14871414_003.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "street", + "choice_b": "restroom", + "choice_c": "car", + "choice_d": "crowded indoor", + "answer_gt": "crowded indoor", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13864 + }, + { + "path": "Elevator_user0684_14977563_006.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "elevator", + "choice_d": "subway station", + "answer_gt": "elevator", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13865 + }, + { + "path": "SubwayStation_user0091_15354072_000.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "elevator", + "choice_b": "street", + "choice_c": "crowded indoor", + "choice_d": "subway station", + "answer_gt": "subway station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13866 + }, + { + "path": "Subway_user0597_14867507_004.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "subway", + "choice_b": "restroom", + "choice_c": "restaurant", + "choice_d": "park", + "answer_gt": "subway", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13867 + }, + { + "path": "Bus_user0330_14875725_003.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "restaurant", + "choice_b": "restroom", + "choice_c": "kitchen", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13868 + }, + { + "path": "Restroom_user0755_14821326_003.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "cafe", + "choice_b": "restaurant", + "choice_c": "restroom", + "choice_d": "crowded indoor", + "answer_gt": "restroom", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13869 + }, + { + "path": "Park_user0024_15018064_004.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "park", + "choice_b": "street", + "choice_c": "restaurant", + "choice_d": "cafe", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13870 + }, + { + "path": "ResidentialArea_user0622_14867959_001.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "restaurant", + "choice_b": "elevator", + "choice_c": "crowded indoor", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "CochlScene", + "uniq_id": 13871 + }, + { + "path": "335.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "forest path", + "choice_b": "metro station", + "choice_c": "bus", + "choice_d": "city center", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13872 + }, + { + "path": "1177.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "office", + "choice_d": "library", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13873 + }, + { + "path": "1546.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13874 + }, + { + "path": "400.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "office", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13875 + }, + { + "path": "529.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "residential area", + "choice_b": "grocery store", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13876 + }, + { + "path": "984.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "library", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13877 + }, + { + "path": "813.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "home", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13878 + }, + { + "path": "497.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "cafe/restaurant", + "choice_b": "library", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13879 + }, + { + "path": "1317.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "office", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13880 + }, + { + "path": "1222.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "library", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13881 + }, + { + "path": "1097.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "library", + "choice_b": "car", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13882 + }, + { + "path": "511.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13883 + }, + { + "path": "1138.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "grocery store", + "choice_b": "forest path", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13884 + }, + { + "path": "303.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "train", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "beach", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13885 + }, + { + "path": "1044.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "car", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13886 + }, + { + "path": "210.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13887 + }, + { + "path": "57.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "residential area", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13888 + }, + { + "path": "895.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "cafe/restaurant", + "choice_b": "office", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13889 + }, + { + "path": "9.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "home", + "choice_d": "beach", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13890 + }, + { + "path": "472.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "office", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13891 + }, + { + "path": "1618.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13892 + }, + { + "path": "711.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13893 + }, + { + "path": "1386.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "city center", + "choice_b": "library", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13894 + }, + { + "path": "33.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13895 + }, + { + "path": "289.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13896 + }, + { + "path": "1345.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "library", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "tram", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13897 + }, + { + "path": "821.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13898 + }, + { + "path": "1040.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13899 + }, + { + "path": "592.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "cafe/restaurant", + "choice_b": "beach", + "choice_c": "tram", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13900 + }, + { + "path": "484.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "home", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13901 + }, + { + "path": "120.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13902 + }, + { + "path": "1155.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "metro station", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13903 + }, + { + "path": "328.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "grocery store", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13904 + }, + { + "path": "1277.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "office", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13905 + }, + { + "path": "1537.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "city center", + "choice_b": "bus", + "choice_c": "tram", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13906 + }, + { + "path": "1581.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "grocery store", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13907 + }, + { + "path": "146.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13908 + }, + { + "path": "1231.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "cafe/restaurant", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13909 + }, + { + "path": "7.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "grocery store", + "choice_b": "home", + "choice_c": "cafe/restaurant", + "choice_d": "residential area", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13910 + }, + { + "path": "1512.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "city center", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13911 + }, + { + "path": "83.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "cafe/restaurant", + "choice_b": "bus", + "choice_c": "city center", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13912 + }, + { + "path": "949.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "home", + "choice_b": "forest path", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13913 + }, + { + "path": "363.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "tram", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13914 + }, + { + "path": "349.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "cafe/restaurant", + "choice_b": "library", + "choice_c": "city center", + "choice_d": "park", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13915 + }, + { + "path": "1568.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "city center", + "choice_b": "bus", + "choice_c": "grocery store", + "choice_d": "car", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13916 + }, + { + "path": "665.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "residential area", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13917 + }, + { + "path": "656.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "office", + "choice_b": "home", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13918 + }, + { + "path": "489.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "car", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13919 + }, + { + "path": "4.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "beach", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13920 + }, + { + "path": "441.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "library", + "choice_b": "grocery store", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13921 + }, + { + "path": "1109.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "library", + "choice_b": "train", + "choice_c": "grocery store", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13922 + }, + { + "path": "1300.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13923 + }, + { + "path": "866.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13924 + }, + { + "path": "1409.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "city center", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13925 + }, + { + "path": "343.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13926 + }, + { + "path": "1046.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13927 + }, + { + "path": "863.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13928 + }, + { + "path": "1410.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13929 + }, + { + "path": "457.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13930 + }, + { + "path": "1163.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "bus", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13931 + }, + { + "path": "1420.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13932 + }, + { + "path": "148.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "city center", + "choice_b": "car", + "choice_c": "home", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13933 + }, + { + "path": "326.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13934 + }, + { + "path": "1185.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "bus", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "park", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13935 + }, + { + "path": "442.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "metro station", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13936 + }, + { + "path": "1171.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "home", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "train", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13937 + }, + { + "path": "344.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "car", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13938 + }, + { + "path": "462.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13939 + }, + { + "path": "855.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13940 + }, + { + "path": "755.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "train", + "choice_b": "office", + "choice_c": "cafe/restaurant", + "choice_d": "home", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13941 + }, + { + "path": "277.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13942 + }, + { + "path": "1020.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "tram", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13943 + }, + { + "path": "756.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "tram", + "choice_b": "park", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13944 + }, + { + "path": "1458.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "train", + "choice_b": "home", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13945 + }, + { + "path": "644.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13946 + }, + { + "path": "624.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "city center", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13947 + }, + { + "path": "1161.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "residential area", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "beach", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13948 + }, + { + "path": "1516.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "grocery store", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13949 + }, + { + "path": "97.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "office", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13950 + }, + { + "path": "1311.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13951 + }, + { + "path": "183.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "park", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13952 + }, + { + "path": "433.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "car", + "choice_b": "city center", + "choice_c": "train", + "choice_d": "office", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13953 + }, + { + "path": "478.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13954 + }, + { + "path": "367.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "office", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13955 + }, + { + "path": "1289.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13956 + }, + { + "path": "508.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "park", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13957 + }, + { + "path": "192.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "office", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13958 + }, + { + "path": "1293.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13959 + }, + { + "path": "568.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "car", + "choice_d": "cafe/restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13960 + }, + { + "path": "498.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "train", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13961 + }, + { + "path": "134.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "cafe/restaurant", + "choice_b": "park", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13962 + }, + { + "path": "913.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "park", + "choice_b": "home", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13963 + }, + { + "path": "1094.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "park", + "choice_d": "tram", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13964 + }, + { + "path": "753.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13965 + }, + { + "path": "573.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13966 + }, + { + "path": "240.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "bus", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13967 + }, + { + "path": "1067.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "metro station", + "choice_d": "park", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13968 + }, + { + "path": "1060.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "city center", + "choice_b": "car", + "choice_c": "home", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13969 + }, + { + "path": "845.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "bus", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13970 + }, + { + "path": "554.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13971 + }, + { + "path": "868.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13972 + }, + { + "path": "610.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13973 + }, + { + "path": "1050.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13974 + }, + { + "path": "617.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "grocery store", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13975 + }, + { + "path": "1614.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13976 + }, + { + "path": "910.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13977 + }, + { + "path": "869.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13978 + }, + { + "path": "1441.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "forest path", + "choice_b": "grocery store", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13979 + }, + { + "path": "10.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13980 + }, + { + "path": "1087.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "train", + "choice_b": "beach", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13981 + }, + { + "path": "1393.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "office", + "choice_b": "home", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13982 + }, + { + "path": "539.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13983 + }, + { + "path": "1114.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "metro station", + "choice_b": "train", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13984 + }, + { + "path": "591.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13985 + }, + { + "path": "228.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "library", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "cafe/restaurant", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13986 + }, + { + "path": "814.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13987 + }, + { + "path": "1327.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13988 + }, + { + "path": "1146.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "tram", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13989 + }, + { + "path": "577.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "forest path", + "choice_d": "bus", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13990 + }, + { + "path": "1106.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13991 + }, + { + "path": "298.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "park", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13992 + }, + { + "path": "917.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "residential area", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13993 + }, + { + "path": "1071.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "tram", + "choice_b": "grocery store", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13994 + }, + { + "path": "773.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "office", + "choice_d": "tram", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13995 + }, + { + "path": "1454.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "grocery store", + "choice_b": "tram", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13996 + }, + { + "path": "679.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13997 + }, + { + "path": "639.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "home", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13998 + }, + { + "path": "357.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 13999 + }, + { + "path": "1256.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "library", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14000 + }, + { + "path": "1365.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14001 + }, + { + "path": "1368.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14002 + }, + { + "path": "66.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "metro station", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14003 + }, + { + "path": "315.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "grocery store", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14004 + }, + { + "path": "260.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "bus", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14005 + }, + { + "path": "1251.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "home", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "beach", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14006 + }, + { + "path": "1276.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14007 + }, + { + "path": "817.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "tram", + "choice_b": "car", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14008 + }, + { + "path": "144.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "park", + "choice_b": "beach", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14009 + }, + { + "path": "708.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14010 + }, + { + "path": "725.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "office", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14011 + }, + { + "path": "257.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "residential area", + "choice_b": "train", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14012 + }, + { + "path": "1396.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "train", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14013 + }, + { + "path": "729.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14014 + }, + { + "path": "1467.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14015 + }, + { + "path": "456.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "tram", + "choice_d": "cafe/restaurant", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14016 + }, + { + "path": "1465.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14017 + }, + { + "path": "734.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "forest path", + "choice_b": "city center", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14018 + }, + { + "path": "107.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "cafe/restaurant", + "choice_b": "residential area", + "choice_c": "grocery store", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14019 + }, + { + "path": "1361.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "library", + "choice_b": "grocery store", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14020 + }, + { + "path": "928.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14021 + }, + { + "path": "1105.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "library", + "choice_b": "metro station", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14022 + }, + { + "path": "1223.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "beach", + "choice_b": "residential area", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14023 + }, + { + "path": "459.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "home", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14024 + }, + { + "path": "412.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14025 + }, + { + "path": "1384.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "tram", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14026 + }, + { + "path": "1574.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "car", + "choice_d": "cafe/restaurant", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14027 + }, + { + "path": "590.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "city center", + "choice_b": "train", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14028 + }, + { + "path": "1174.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "park", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14029 + }, + { + "path": "652.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "tram", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14030 + }, + { + "path": "744.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "car", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14031 + }, + { + "path": "26.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "grocery store", + "choice_d": "beach", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14032 + }, + { + "path": "635.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14033 + }, + { + "path": "1047.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14034 + }, + { + "path": "404.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14035 + }, + { + "path": "1474.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14036 + }, + { + "path": "854.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "city center", + "choice_b": "train", + "choice_c": "home", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14037 + }, + { + "path": "1325.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14038 + }, + { + "path": "248.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14039 + }, + { + "path": "1423.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "city center", + "choice_b": "forest path", + "choice_c": "bus", + "choice_d": "library", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14040 + }, + { + "path": "633.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "home", + "choice_b": "city center", + "choice_c": "library", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14041 + }, + { + "path": "1125.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "cafe/restaurant", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "bus", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14042 + }, + { + "path": "200.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "tram", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14043 + }, + { + "path": "124.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14044 + }, + { + "path": "758.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14045 + }, + { + "path": "68.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14046 + }, + { + "path": "513.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "car", + "choice_b": "grocery store", + "choice_c": "office", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14047 + }, + { + "path": "1142.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "library", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14048 + }, + { + "path": "953.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14049 + }, + { + "path": "579.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "library", + "choice_b": "tram", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14050 + }, + { + "path": "1422.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "cafe/restaurant", + "choice_d": "home", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14051 + }, + { + "path": "1457.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14052 + }, + { + "path": "776.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "beach", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14053 + }, + { + "path": "615.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "beach", + "choice_b": "grocery store", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14054 + }, + { + "path": "98.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14055 + }, + { + "path": "560.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "grocery store", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14056 + }, + { + "path": "1079.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14057 + }, + { + "path": "1149.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "metro station", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14058 + }, + { + "path": "1493.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "metro station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14059 + }, + { + "path": "222.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "forest path", + "choice_b": "grocery store", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14060 + }, + { + "path": "155.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14061 + }, + { + "path": "1338.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14062 + }, + { + "path": "1252.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "office", + "choice_d": "city center", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14063 + }, + { + "path": "1434.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "library", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14064 + }, + { + "path": "1190.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "train", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "office", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14065 + }, + { + "path": "1029.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14066 + }, + { + "path": "975.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14067 + }, + { + "path": "648.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "library", + "choice_d": "home", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14068 + }, + { + "path": "1585.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "forest path", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14069 + }, + { + "path": "1246.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14070 + }, + { + "path": "779.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14071 + }, + { + "path": "1503.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "residential area", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14072 + }, + { + "path": "171.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "train", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "cafe/restaurant", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14073 + }, + { + "path": "1080.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "cafe/restaurant", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14074 + }, + { + "path": "1456.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "city center", + "choice_b": "beach", + "choice_c": "cafe/restaurant", + "choice_d": "tram", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14075 + }, + { + "path": "704.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14076 + }, + { + "path": "1266.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "residential area", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14077 + }, + { + "path": "1226.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "residential area", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14078 + }, + { + "path": "440.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "park", + "choice_b": "forest path", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14079 + }, + { + "path": "1153.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "home", + "choice_b": "beach", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14080 + }, + { + "path": "1202.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "office", + "choice_d": "cafe/restaurant", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14081 + }, + { + "path": "681.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14082 + }, + { + "path": "1038.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "metro station", + "choice_b": "beach", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14083 + }, + { + "path": "1479.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "metro station", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14084 + }, + { + "path": "1520.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14085 + }, + { + "path": "1150.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "tram", + "choice_b": "car", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14086 + }, + { + "path": "259.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14087 + }, + { + "path": "52.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "car", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14088 + }, + { + "path": "417.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14089 + }, + { + "path": "1371.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "home", + "choice_d": "forest path", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14090 + }, + { + "path": "1282.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14091 + }, + { + "path": "520.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "metro station", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14092 + }, + { + "path": "390.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14093 + }, + { + "path": "361.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14094 + }, + { + "path": "1405.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "metro station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14095 + }, + { + "path": "718.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "grocery store", + "choice_b": "forest path", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14096 + }, + { + "path": "789.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "office", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14097 + }, + { + "path": "1543.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "tram", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14098 + }, + { + "path": "1294.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "tram", + "choice_b": "office", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14099 + }, + { + "path": "673.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "office", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "forest path", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14100 + }, + { + "path": "738.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14101 + }, + { + "path": "1262.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14102 + }, + { + "path": "1245.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "beach", + "choice_d": "office", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14103 + }, + { + "path": "995.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14104 + }, + { + "path": "1103.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14105 + }, + { + "path": "1314.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14106 + }, + { + "path": "1243.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14107 + }, + { + "path": "261.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "forest path", + "choice_b": "metro station", + "choice_c": "library", + "choice_d": "grocery store", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14108 + }, + { + "path": "959.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "library", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14109 + }, + { + "path": "382.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "bus", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14110 + }, + { + "path": "39.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "car", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14111 + }, + { + "path": "1435.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14112 + }, + { + "path": "150.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "home", + "choice_d": "forest path", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14113 + }, + { + "path": "111.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "cafe/restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14114 + }, + { + "path": "1406.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "office", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14115 + }, + { + "path": "118.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "residential area", + "choice_b": "cafe/restaurant", + "choice_c": "city center", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14116 + }, + { + "path": "827.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14117 + }, + { + "path": "1578.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14118 + }, + { + "path": "1237.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14119 + }, + { + "path": "487.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14120 + }, + { + "path": "1053.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14121 + }, + { + "path": "467.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14122 + }, + { + "path": "1025.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "grocery store", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14123 + }, + { + "path": "362.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14124 + }, + { + "path": "1209.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14125 + }, + { + "path": "569.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "home", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14126 + }, + { + "path": "908.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "car", + "choice_d": "beach", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14127 + }, + { + "path": "637.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14128 + }, + { + "path": "385.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "metro station", + "choice_b": "city center", + "choice_c": "train", + "choice_d": "grocery store", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14129 + }, + { + "path": "856.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "beach", + "choice_b": "metro station", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14130 + }, + { + "path": "1525.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14131 + }, + { + "path": "631.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "office", + "choice_b": "tram", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14132 + }, + { + "path": "878.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "grocery store", + "choice_b": "residential area", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14133 + }, + { + "path": "1269.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "home", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "tram", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14134 + }, + { + "path": "355.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "train", + "choice_b": "home", + "choice_c": "library", + "choice_d": "forest path", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14135 + }, + { + "path": "836.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "library", + "choice_b": "residential area", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14136 + }, + { + "path": "330.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "beach", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14137 + }, + { + "path": "745.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "city center", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14138 + }, + { + "path": "927.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "city center", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14139 + }, + { + "path": "881.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "beach", + "choice_c": "residential area", + "choice_d": "home", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14140 + }, + { + "path": "273.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "forest path", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14141 + }, + { + "path": "1342.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14142 + }, + { + "path": "1225.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14143 + }, + { + "path": "1274.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14144 + }, + { + "path": "112.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14145 + }, + { + "path": "56.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "grocery store", + "choice_b": "cafe/restaurant", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14146 + }, + { + "path": "919.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14147 + }, + { + "path": "89.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "tram", + "choice_b": "grocery store", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14148 + }, + { + "path": "674.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "cafe/restaurant", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14149 + }, + { + "path": "1054.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "train", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14150 + }, + { + "path": "1204.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14151 + }, + { + "path": "1275.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "bus", + "choice_d": "city center", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14152 + }, + { + "path": "819.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14153 + }, + { + "path": "143.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "grocery store", + "choice_b": "library", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14154 + }, + { + "path": "1447.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14155 + }, + { + "path": "1157.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14156 + }, + { + "path": "1364.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14157 + }, + { + "path": "126.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "office", + "choice_b": "car", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14158 + }, + { + "path": "1489.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "office", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14159 + }, + { + "path": "299.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "city center", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14160 + }, + { + "path": "420.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe/restaurant", + "choice_b": "grocery store", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14161 + }, + { + "path": "1188.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "cafe/restaurant", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14162 + }, + { + "path": "985.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "train", + "choice_b": "beach", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14163 + }, + { + "path": "1255.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "office", + "choice_b": "beach", + "choice_c": "library", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14164 + }, + { + "path": "1005.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "forest path", + "choice_d": "library", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14165 + }, + { + "path": "546.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "park", + "choice_d": "tram", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14166 + }, + { + "path": "1332.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14167 + }, + { + "path": "1597.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14168 + }, + { + "path": "1475.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14169 + }, + { + "path": "1219.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "car", + "choice_b": "city center", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14170 + }, + { + "path": "930.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "bus", + "choice_b": "forest path", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14171 + }, + { + "path": "896.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14172 + }, + { + "path": "1090.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "cafe/restaurant", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14173 + }, + { + "path": "309.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "car", + "choice_b": "city center", + "choice_c": "park", + "choice_d": "forest path", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14174 + }, + { + "path": "561.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14175 + }, + { + "path": "1102.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "train", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14176 + }, + { + "path": "1192.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14177 + }, + { + "path": "696.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "office", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14178 + }, + { + "path": "778.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "metro station", + "choice_b": "train", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14179 + }, + { + "path": "1213.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "cafe/restaurant", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14180 + }, + { + "path": "1519.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "office", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14181 + }, + { + "path": "1545.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14182 + }, + { + "path": "283.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14183 + }, + { + "path": "1544.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "bus", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "grocery store", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14184 + }, + { + "path": "532.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14185 + }, + { + "path": "1013.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "residential area", + "choice_b": "grocery store", + "choice_c": "beach", + "choice_d": "office", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14186 + }, + { + "path": "1558.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14187 + }, + { + "path": "667.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14188 + }, + { + "path": "445.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "grocery store", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14189 + }, + { + "path": "95.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "grocery store", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14190 + }, + { + "path": "611.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "city center", + "choice_b": "library", + "choice_c": "cafe/restaurant", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14191 + }, + { + "path": "1201.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14192 + }, + { + "path": "1528.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14193 + }, + { + "path": "91.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "office", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "library", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14194 + }, + { + "path": "1297.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "train", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14195 + }, + { + "path": "1176.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "residential area", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14196 + }, + { + "path": "1460.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "park", + "choice_b": "grocery store", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14197 + }, + { + "path": "1370.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14198 + }, + { + "path": "646.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "library", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "beach", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14199 + }, + { + "path": "1230.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14200 + }, + { + "path": "11.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "cafe/restaurant", + "choice_b": "office", + "choice_c": "library", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14201 + }, + { + "path": "609.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "beach", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14202 + }, + { + "path": "78.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14203 + }, + { + "path": "942.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "forest path", + "choice_b": "car", + "choice_c": "office", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14204 + }, + { + "path": "1260.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14205 + }, + { + "path": "1295.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "grocery store", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14206 + }, + { + "path": "503.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "cafe/restaurant", + "choice_b": "bus", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14207 + }, + { + "path": "1248.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "train", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14208 + }, + { + "path": "1514.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "train", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14209 + }, + { + "path": "49.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "city center", + "choice_b": "grocery store", + "choice_c": "office", + "choice_d": "bus", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14210 + }, + { + "path": "1008.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14211 + }, + { + "path": "1181.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "city center", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "residential area", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14212 + }, + { + "path": "873.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14213 + }, + { + "path": "702.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "tram", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14214 + }, + { + "path": "882.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14215 + }, + { + "path": "329.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "tram", + "choice_b": "library", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14216 + }, + { + "path": "597.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14217 + }, + { + "path": "943.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "office", + "choice_b": "car", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14218 + }, + { + "path": "1261.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "home", + "choice_b": "beach", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14219 + }, + { + "path": "749.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "residential area", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14220 + }, + { + "path": "292.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "tram", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "library", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14221 + }, + { + "path": "922.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "park", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14222 + }, + { + "path": "1387.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14223 + }, + { + "path": "843.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14224 + }, + { + "path": "163.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14225 + }, + { + "path": "1221.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "metro station", + "choice_d": "library", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14226 + }, + { + "path": "399.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "residential area", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14227 + }, + { + "path": "751.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "tram", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14228 + }, + { + "path": "47.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "city center", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14229 + }, + { + "path": "1306.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14230 + }, + { + "path": "526.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14231 + }, + { + "path": "1081.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "train", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14232 + }, + { + "path": "980.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14233 + }, + { + "path": "969.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14234 + }, + { + "path": "473.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "tram", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14235 + }, + { + "path": "368.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "office", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14236 + }, + { + "path": "226.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "grocery store", + "choice_b": "library", + "choice_c": "office", + "choice_d": "city center", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14237 + }, + { + "path": "1272.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "forest path", + "choice_b": "car", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14238 + }, + { + "path": "782.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "cafe/restaurant", + "choice_d": "bus", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14239 + }, + { + "path": "1596.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14240 + }, + { + "path": "551.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "grocery store", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "tram", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14241 + }, + { + "path": "465.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "home", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14242 + }, + { + "path": "721.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14243 + }, + { + "path": "857.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "car", + "choice_b": "cafe/restaurant", + "choice_c": "residential area", + "choice_d": "forest path", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14244 + }, + { + "path": "1494.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "forest path", + "choice_b": "train", + "choice_c": "office", + "choice_d": "car", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14245 + }, + { + "path": "1355.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "grocery store", + "choice_b": "bus", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14246 + }, + { + "path": "1236.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "grocery store", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14247 + }, + { + "path": "1402.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "grocery store", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14248 + }, + { + "path": "177.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14249 + }, + { + "path": "500.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "grocery store", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "office", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14250 + }, + { + "path": "1445.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "cafe/restaurant", + "choice_b": "home", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14251 + }, + { + "path": "1397.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "grocery store", + "choice_d": "residential area", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14252 + }, + { + "path": "785.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14253 + }, + { + "path": "746.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "park", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14254 + }, + { + "path": "1208.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "grocery store", + "choice_b": "train", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14255 + }, + { + "path": "1122.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "city center", + "choice_d": "grocery store", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14256 + }, + { + "path": "199.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "library", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14257 + }, + { + "path": "444.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "office", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14258 + }, + { + "path": "99.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "home", + "choice_b": "grocery store", + "choice_c": "beach", + "choice_d": "bus", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14259 + }, + { + "path": "419.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "cafe/restaurant", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14260 + }, + { + "path": "1484.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "bus", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14261 + }, + { + "path": "1113.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14262 + }, + { + "path": "247.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14263 + }, + { + "path": "1093.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14264 + }, + { + "path": "211.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "office", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14265 + }, + { + "path": "759.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "park", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14266 + }, + { + "path": "488.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "library", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14267 + }, + { + "path": "1211.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "train", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14268 + }, + { + "path": "249.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "home", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14269 + }, + { + "path": "621.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14270 + }, + { + "path": "530.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14271 + }, + { + "path": "231.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "library", + "choice_b": "cafe/restaurant", + "choice_c": "beach", + "choice_d": "residential area", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14272 + }, + { + "path": "1316.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14273 + }, + { + "path": "740.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14274 + }, + { + "path": "1459.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "library", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14275 + }, + { + "path": "1257.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "grocery store", + "choice_b": "home", + "choice_c": "forest path", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14276 + }, + { + "path": "1417.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14277 + }, + { + "path": "559.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "home", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14278 + }, + { + "path": "1066.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "train", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14279 + }, + { + "path": "692.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14280 + }, + { + "path": "1078.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14281 + }, + { + "path": "892.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "library", + "choice_d": "forest path", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14282 + }, + { + "path": "1124.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "grocery store", + "choice_b": "office", + "choice_c": "library", + "choice_d": "home", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14283 + }, + { + "path": "733.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14284 + }, + { + "path": "1035.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14285 + }, + { + "path": "1267.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "library", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14286 + }, + { + "path": "1584.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14287 + }, + { + "path": "1572.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14288 + }, + { + "path": "1195.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "metro station", + "choice_b": "city center", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14289 + }, + { + "path": "1601.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "office", + "choice_b": "beach", + "choice_c": "grocery store", + "choice_d": "car", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14290 + }, + { + "path": "811.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "home", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14291 + }, + { + "path": "130.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "forest path", + "choice_b": "library", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14292 + }, + { + "path": "1617.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14293 + }, + { + "path": "1412.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "library", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14294 + }, + { + "path": "1009.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "train", + "choice_b": "grocery store", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14295 + }, + { + "path": "1207.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "metro station", + "choice_d": "office", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14296 + }, + { + "path": "564.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14297 + }, + { + "path": "1159.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14298 + }, + { + "path": "587.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "car", + "choice_b": "metro station", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14299 + }, + { + "path": "274.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14300 + }, + { + "path": "979.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "cafe/restaurant", + "choice_d": "residential area", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14301 + }, + { + "path": "270.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "residential area", + "choice_b": "forest path", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14302 + }, + { + "path": "525.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "train", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14303 + }, + { + "path": "1214.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "city center", + "choice_b": "park", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14304 + }, + { + "path": "1576.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14305 + }, + { + "path": "517.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14306 + }, + { + "path": "258.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "city center", + "choice_b": "train", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14307 + }, + { + "path": "588.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "grocery store", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14308 + }, + { + "path": "23.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "library", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14309 + }, + { + "path": "1589.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "office", + "choice_b": "city center", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14310 + }, + { + "path": "301.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "residential area", + "choice_b": "cafe/restaurant", + "choice_c": "home", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14311 + }, + { + "path": "221.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "city center", + "choice_b": "car", + "choice_c": "beach", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14312 + }, + { + "path": "1440.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14313 + }, + { + "path": "800.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14314 + }, + { + "path": "1028.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "bus", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14315 + }, + { + "path": "657.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "cafe/restaurant", + "choice_d": "home", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14316 + }, + { + "path": "350.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "train", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14317 + }, + { + "path": "250.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "library", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14318 + }, + { + "path": "22.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "office", + "choice_d": "grocery store", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14319 + }, + { + "path": "1605.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14320 + }, + { + "path": "271.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "cafe/restaurant", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "forest path", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14321 + }, + { + "path": "994.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "park", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14322 + }, + { + "path": "1594.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "library", + "choice_b": "tram", + "choice_c": "metro station", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14323 + }, + { + "path": "1074.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "beach", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14324 + }, + { + "path": "904.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14325 + }, + { + "path": "72.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "forest path", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14326 + }, + { + "path": "1547.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "residential area", + "choice_b": "home", + "choice_c": "office", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14327 + }, + { + "path": "605.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "office", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "cafe/restaurant", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14328 + }, + { + "path": "629.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "city center", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "library", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14329 + }, + { + "path": "1001.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14330 + }, + { + "path": "67.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "beach", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14331 + }, + { + "path": "322.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "office", + "choice_d": "city center", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14332 + }, + { + "path": "114.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "city center", + "choice_d": "office", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14333 + }, + { + "path": "804.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "city center", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14334 + }, + { + "path": "1376.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14335 + }, + { + "path": "493.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "home", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14336 + }, + { + "path": "1198.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "office", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14337 + }, + { + "path": "61.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14338 + }, + { + "path": "867.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "library", + "choice_b": "beach", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14339 + }, + { + "path": "699.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "grocery store", + "choice_b": "library", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14340 + }, + { + "path": "481.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14341 + }, + { + "path": "1553.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14342 + }, + { + "path": "1117.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "forest path", + "choice_b": "metro station", + "choice_c": "office", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14343 + }, + { + "path": "849.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14344 + }, + { + "path": "395.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14345 + }, + { + "path": "108.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "home", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14346 + }, + { + "path": "466.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "library", + "choice_b": "beach", + "choice_c": "city center", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14347 + }, + { + "path": "1497.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14348 + }, + { + "path": "387.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "home", + "choice_b": "city center", + "choice_c": "forest path", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14349 + }, + { + "path": "1085.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "tram", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "train", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14350 + }, + { + "path": "709.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14351 + }, + { + "path": "893.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "home", + "choice_b": "metro station", + "choice_c": "city center", + "choice_d": "train", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14352 + }, + { + "path": "149.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "cafe/restaurant", + "choice_b": "park", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14353 + }, + { + "path": "1193.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "grocery store", + "choice_b": "forest path", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14354 + }, + { + "path": "1145.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14355 + }, + { + "path": "312.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "grocery store", + "choice_b": "forest path", + "choice_c": "cafe/restaurant", + "choice_d": "car", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14356 + }, + { + "path": "1401.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "office", + "choice_d": "bus", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14357 + }, + { + "path": "1043.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "library", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "forest path", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14358 + }, + { + "path": "750.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "car", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14359 + }, + { + "path": "374.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "train", + "choice_d": "residential area", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14360 + }, + { + "path": "557.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14361 + }, + { + "path": "1346.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14362 + }, + { + "path": "1129.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14363 + }, + { + "path": "169.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "bus", + "choice_b": "tram", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14364 + }, + { + "path": "1472.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "library", + "choice_d": "residential area", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14365 + }, + { + "path": "370.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14366 + }, + { + "path": "1110.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14367 + }, + { + "path": "376.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14368 + }, + { + "path": "965.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "car", + "choice_b": "beach", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14369 + }, + { + "path": "1517.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14370 + }, + { + "path": "461.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14371 + }, + { + "path": "575.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14372 + }, + { + "path": "1350.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "grocery store", + "choice_b": "park", + "choice_c": "forest path", + "choice_d": "beach", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14373 + }, + { + "path": "894.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "beach", + "choice_b": "train", + "choice_c": "library", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14374 + }, + { + "path": "1427.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "beach", + "choice_b": "office", + "choice_c": "park", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14375 + }, + { + "path": "1203.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "city center", + "choice_b": "home", + "choice_c": "residential area", + "choice_d": "office", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14376 + }, + { + "path": "446.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "bus", + "choice_b": "metro station", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14377 + }, + { + "path": "1165.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14378 + }, + { + "path": "1286.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "office", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "home", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14379 + }, + { + "path": "1439.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14380 + }, + { + "path": "1.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14381 + }, + { + "path": "1318.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "home", + "choice_b": "city center", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14382 + }, + { + "path": "19.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14383 + }, + { + "path": "1033.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "city center", + "choice_b": "car", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14384 + }, + { + "path": "1126.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14385 + }, + { + "path": "65.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "library", + "choice_b": "city center", + "choice_c": "cafe/restaurant", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14386 + }, + { + "path": "492.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "office", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14387 + }, + { + "path": "932.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "forest path", + "choice_b": "city center", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14388 + }, + { + "path": "790.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14389 + }, + { + "path": "636.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "park", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14390 + }, + { + "path": "732.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "cafe/restaurant", + "choice_b": "office", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14391 + }, + { + "path": "1280.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14392 + }, + { + "path": "1557.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "car", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14393 + }, + { + "path": "486.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "office", + "choice_d": "car", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14394 + }, + { + "path": "116.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "metro station", + "choice_b": "grocery store", + "choice_c": "car", + "choice_d": "cafe/restaurant", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14395 + }, + { + "path": "423.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14396 + }, + { + "path": "874.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "residential area", + "choice_d": "library", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14397 + }, + { + "path": "1343.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14398 + }, + { + "path": "987.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14399 + }, + { + "path": "805.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "park", + "choice_d": "cafe/restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14400 + }, + { + "path": "754.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "grocery store", + "choice_b": "home", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14401 + }, + { + "path": "1075.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "home", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14402 + }, + { + "path": "1116.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14403 + }, + { + "path": "219.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "city center", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14404 + }, + { + "path": "409.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "grocery store", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14405 + }, + { + "path": "244.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14406 + }, + { + "path": "936.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "grocery store", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14407 + }, + { + "path": "1366.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "metro station", + "choice_b": "beach", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14408 + }, + { + "path": "321.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "park", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14409 + }, + { + "path": "356.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "forest path", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14410 + }, + { + "path": "1104.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "tram", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14411 + }, + { + "path": "208.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "grocery store", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14412 + }, + { + "path": "1303.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "library", + "choice_d": "home", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14413 + }, + { + "path": "940.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14414 + }, + { + "path": "742.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "park", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14415 + }, + { + "path": "622.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "train", + "choice_d": "grocery store", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14416 + }, + { + "path": "1510.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "residential area", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14417 + }, + { + "path": "280.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14418 + }, + { + "path": "479.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14419 + }, + { + "path": "131.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "park", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14420 + }, + { + "path": "660.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14421 + }, + { + "path": "87.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "cafe/restaurant", + "choice_b": "tram", + "choice_c": "forest path", + "choice_d": "bus", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14422 + }, + { + "path": "485.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14423 + }, + { + "path": "1486.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "library", + "choice_d": "train", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14424 + }, + { + "path": "331.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14425 + }, + { + "path": "165.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "office", + "choice_b": "grocery store", + "choice_c": "metro station", + "choice_d": "bus", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14426 + }, + { + "path": "386.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14427 + }, + { + "path": "888.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "forest path", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14428 + }, + { + "path": "1414.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "city center", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14429 + }, + { + "path": "1561.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14430 + }, + { + "path": "599.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "train", + "choice_b": "park", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14431 + }, + { + "path": "251.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14432 + }, + { + "path": "921.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14433 + }, + { + "path": "906.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "car", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "bus", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14434 + }, + { + "path": "69.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "forest path", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14435 + }, + { + "path": "64.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "train", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14436 + }, + { + "path": "1045.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "office", + "choice_d": "library", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14437 + }, + { + "path": "241.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "residential area", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14438 + }, + { + "path": "562.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "cafe/restaurant", + "choice_b": "metro station", + "choice_c": "library", + "choice_d": "home", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14439 + }, + { + "path": "1301.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14440 + }, + { + "path": "1513.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14441 + }, + { + "path": "193.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14442 + }, + { + "path": "889.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14443 + }, + { + "path": "886.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe/restaurant", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "beach", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14444 + }, + { + "path": "997.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "car", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14445 + }, + { + "path": "406.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "cafe/restaurant", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14446 + }, + { + "path": "297.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14447 + }, + { + "path": "1095.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "home", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14448 + }, + { + "path": "53.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "park", + "choice_b": "beach", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14449 + }, + { + "path": "1403.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "home", + "choice_b": "metro station", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14450 + }, + { + "path": "833.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14451 + }, + { + "path": "311.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "office", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14452 + }, + { + "path": "187.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14453 + }, + { + "path": "1524.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "car", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14454 + }, + { + "path": "1582.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "beach", + "choice_b": "residential area", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14455 + }, + { + "path": "852.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14456 + }, + { + "path": "1315.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14457 + }, + { + "path": "1014.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "car", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14458 + }, + { + "path": "142.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14459 + }, + { + "path": "34.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "forest path", + "choice_b": "city center", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14460 + }, + { + "path": "682.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "beach", + "choice_b": "city center", + "choice_c": "train", + "choice_d": "residential area", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14461 + }, + { + "path": "1002.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "train", + "choice_b": "beach", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14462 + }, + { + "path": "1136.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "tram", + "choice_c": "beach", + "choice_d": "car", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14463 + }, + { + "path": "807.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "forest path", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14464 + }, + { + "path": "731.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "tram", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14465 + }, + { + "path": "558.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14466 + }, + { + "path": "471.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "home", + "choice_d": "city center", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14467 + }, + { + "path": "1426.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14468 + }, + { + "path": "46.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "beach", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14469 + }, + { + "path": "608.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14470 + }, + { + "path": "403.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14471 + }, + { + "path": "1468.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "cafe/restaurant", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14472 + }, + { + "path": "1395.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "train", + "choice_b": "grocery store", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14473 + }, + { + "path": "726.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14474 + }, + { + "path": "18.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "residential area", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14475 + }, + { + "path": "58.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "forest path", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14476 + }, + { + "path": "346.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "bus", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "home", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14477 + }, + { + "path": "694.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "library", + "choice_b": "train", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14478 + }, + { + "path": "428.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "library", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14479 + }, + { + "path": "1483.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14480 + }, + { + "path": "595.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14481 + }, + { + "path": "879.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14482 + }, + { + "path": "697.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "bus", + "choice_b": "city center", + "choice_c": "grocery store", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14483 + }, + { + "path": "268.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "bus", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14484 + }, + { + "path": "507.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "city center", + "choice_b": "office", + "choice_c": "park", + "choice_d": "cafe/restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14485 + }, + { + "path": "1374.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "train", + "choice_b": "cafe/restaurant", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14486 + }, + { + "path": "339.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "library", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14487 + }, + { + "path": "802.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "grocery store", + "choice_b": "park", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14488 + }, + { + "path": "1228.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "library", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "home", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14489 + }, + { + "path": "527.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "bus", + "choice_b": "library", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14490 + }, + { + "path": "1391.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14491 + }, + { + "path": "449.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "grocery store", + "choice_b": "train", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14492 + }, + { + "path": "524.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14493 + }, + { + "path": "929.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "library", + "choice_b": "park", + "choice_c": "car", + "choice_d": "cafe/restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14494 + }, + { + "path": "661.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14495 + }, + { + "path": "1299.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14496 + }, + { + "path": "831.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "home", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14497 + }, + { + "path": "848.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "train", + "choice_b": "city center", + "choice_c": "grocery store", + "choice_d": "tram", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14498 + }, + { + "path": "722.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14499 + }, + { + "path": "1139.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "office", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14500 + }, + { + "path": "1565.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "cafe/restaurant", + "choice_b": "residential area", + "choice_c": "grocery store", + "choice_d": "city center", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14501 + }, + { + "path": "63.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "library", + "choice_b": "bus", + "choice_c": "tram", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14502 + }, + { + "path": "1551.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14503 + }, + { + "path": "1351.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "metro station", + "choice_b": "beach", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14504 + }, + { + "path": "28.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "beach", + "choice_b": "grocery store", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14505 + }, + { + "path": "1285.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "beach", + "choice_d": "tram", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14506 + }, + { + "path": "1175.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "car", + "choice_b": "office", + "choice_c": "grocery store", + "choice_d": "cafe/restaurant", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14507 + }, + { + "path": "40.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14508 + }, + { + "path": "543.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "office", + "choice_d": "bus", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14509 + }, + { + "path": "1538.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "library", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14510 + }, + { + "path": "1579.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "city center", + "choice_b": "home", + "choice_c": "metro station", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14511 + }, + { + "path": "1302.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "tram", + "choice_b": "forest path", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14512 + }, + { + "path": "1362.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "beach", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14513 + }, + { + "path": "388.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14514 + }, + { + "path": "1115.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "cafe/restaurant", + "choice_b": "library", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14515 + }, + { + "path": "640.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14516 + }, + { + "path": "431.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "office", + "choice_b": "metro station", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14517 + }, + { + "path": "825.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "tram", + "choice_b": "grocery store", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14518 + }, + { + "path": "992.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "car", + "choice_d": "library", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14519 + }, + { + "path": "623.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "park", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14520 + }, + { + "path": "523.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "forest path", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14521 + }, + { + "path": "201.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14522 + }, + { + "path": "377.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14523 + }, + { + "path": "1437.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "bus", + "choice_b": "park", + "choice_c": "beach", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14524 + }, + { + "path": "900.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "train", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14525 + }, + { + "path": "392.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "home", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14526 + }, + { + "path": "1329.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "train", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14527 + }, + { + "path": "1059.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "city center", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14528 + }, + { + "path": "1023.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "grocery store", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14529 + }, + { + "path": "215.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "grocery store", + "choice_b": "tram", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14530 + }, + { + "path": "291.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "grocery store", + "choice_c": "home", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14531 + }, + { + "path": "1027.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "home", + "choice_b": "cafe/restaurant", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14532 + }, + { + "path": "982.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14533 + }, + { + "path": "325.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14534 + }, + { + "path": "115.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "beach", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14535 + }, + { + "path": "1323.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "park", + "choice_b": "train", + "choice_c": "car", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14536 + }, + { + "path": "757.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14537 + }, + { + "path": "398.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "train", + "choice_b": "cafe/restaurant", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14538 + }, + { + "path": "48.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "park", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14539 + }, + { + "path": "245.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14540 + }, + { + "path": "510.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "library", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14541 + }, + { + "path": "188.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "library", + "choice_b": "forest path", + "choice_c": "grocery store", + "choice_d": "tram", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14542 + }, + { + "path": "1281.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe/restaurant", + "choice_b": "metro station", + "choice_c": "forest path", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14543 + }, + { + "path": "565.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "residential area", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14544 + }, + { + "path": "1542.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14545 + }, + { + "path": "405.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "car", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "city center", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14546 + }, + { + "path": "705.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "metro station", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14547 + }, + { + "path": "225.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "office", + "choice_d": "train", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14548 + }, + { + "path": "31.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "train", + "choice_b": "grocery store", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14549 + }, + { + "path": "555.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14550 + }, + { + "path": "207.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "cafe/restaurant", + "choice_d": "tram", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14551 + }, + { + "path": "418.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "park", + "choice_b": "train", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14552 + }, + { + "path": "424.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "tram", + "choice_b": "car", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14553 + }, + { + "path": "264.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14554 + }, + { + "path": "128.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "tram", + "choice_d": "beach", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14555 + }, + { + "path": "27.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "tram", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14556 + }, + { + "path": "1606.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "train", + "choice_d": "car", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14557 + }, + { + "path": "1431.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "park", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14558 + }, + { + "path": "173.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "bus", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14559 + }, + { + "path": "430.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "forest path", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14560 + }, + { + "path": "1591.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "beach", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14561 + }, + { + "path": "1344.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "car", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14562 + }, + { + "path": "810.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "tram", + "choice_b": "home", + "choice_c": "park", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14563 + }, + { + "path": "1187.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "home", + "choice_d": "metro station", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14564 + }, + { + "path": "1006.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "office", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14565 + }, + { + "path": "926.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "grocery store", + "choice_b": "home", + "choice_c": "metro station", + "choice_d": "residential area", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14566 + }, + { + "path": "262.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "tram", + "choice_b": "metro station", + "choice_c": "cafe/restaurant", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14567 + }, + { + "path": "1127.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "tram", + "choice_b": "park", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14568 + }, + { + "path": "1587.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "home", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14569 + }, + { + "path": "752.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14570 + }, + { + "path": "1488.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "car", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14571 + }, + { + "path": "8.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "city center", + "choice_d": "residential area", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14572 + }, + { + "path": "1271.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "metro station", + "choice_b": "library", + "choice_c": "cafe/restaurant", + "choice_d": "beach", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14573 + }, + { + "path": "1612.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "metro station", + "choice_b": "forest path", + "choice_c": "grocery store", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14574 + }, + { + "path": "1382.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "forest path", + "choice_b": "home", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14575 + }, + { + "path": "1052.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14576 + }, + { + "path": "1119.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "tram", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14577 + }, + { + "path": "771.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "park", + "choice_b": "beach", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14578 + }, + { + "path": "838.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "library", + "choice_b": "cafe/restaurant", + "choice_c": "beach", + "choice_d": "tram", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14579 + }, + { + "path": "1101.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "residential area", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14580 + }, + { + "path": "1416.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "beach", + "choice_b": "grocery store", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14581 + }, + { + "path": "1619.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14582 + }, + { + "path": "850.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14583 + }, + { + "path": "438.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "city center", + "choice_b": "tram", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14584 + }, + { + "path": "1462.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "forest path", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14585 + }, + { + "path": "944.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14586 + }, + { + "path": "872.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14587 + }, + { + "path": "81.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "library", + "choice_b": "forest path", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14588 + }, + { + "path": "1529.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "office", + "choice_b": "grocery store", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14589 + }, + { + "path": "1577.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "train", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14590 + }, + { + "path": "246.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "bus", + "choice_b": "forest path", + "choice_c": "city center", + "choice_d": "car", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14591 + }, + { + "path": "1241.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14592 + }, + { + "path": "781.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "library", + "choice_b": "beach", + "choice_c": "car", + "choice_d": "train", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14593 + }, + { + "path": "338.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "park", + "choice_b": "metro station", + "choice_c": "office", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14594 + }, + { + "path": "570.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "grocery store", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14595 + }, + { + "path": "20.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "office", + "choice_d": "cafe/restaurant", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14596 + }, + { + "path": "841.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14597 + }, + { + "path": "865.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14598 + }, + { + "path": "616.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14599 + }, + { + "path": "76.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "forest path", + "choice_b": "car", + "choice_c": "grocery store", + "choice_d": "office", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14600 + }, + { + "path": "514.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "bus", + "choice_b": "residential area", + "choice_c": "forest path", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14601 + }, + { + "path": "1034.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "office", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14602 + }, + { + "path": "712.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "tram", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14603 + }, + { + "path": "1258.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "residential area", + "choice_b": "car", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14604 + }, + { + "path": "1404.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "residential area", + "choice_b": "train", + "choice_c": "metro station", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14605 + }, + { + "path": "535.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "residential area", + "choice_b": "home", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14606 + }, + { + "path": "94.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "park", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14607 + }, + { + "path": "287.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14608 + }, + { + "path": "536.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "car", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "city center", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14609 + }, + { + "path": "6.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "office", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14610 + }, + { + "path": "1194.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14611 + }, + { + "path": "214.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "park", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14612 + }, + { + "path": "870.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14613 + }, + { + "path": "179.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "cafe/restaurant", + "choice_b": "city center", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14614 + }, + { + "path": "1598.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14615 + }, + { + "path": "567.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "city center", + "choice_b": "train", + "choice_c": "forest path", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14616 + }, + { + "path": "1433.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14617 + }, + { + "path": "939.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "car", + "choice_b": "train", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14618 + }, + { + "path": "42.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "bus", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14619 + }, + { + "path": "1234.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "residential area", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14620 + }, + { + "path": "1143.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "tram", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14621 + }, + { + "path": "1442.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14622 + }, + { + "path": "109.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "beach", + "choice_b": "city center", + "choice_c": "metro station", + "choice_d": "library", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14623 + }, + { + "path": "1212.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "car", + "choice_b": "tram", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14624 + }, + { + "path": "830.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "library", + "choice_d": "metro station", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14625 + }, + { + "path": "1476.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "cafe/restaurant", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14626 + }, + { + "path": "736.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "office", + "choice_b": "forest path", + "choice_c": "tram", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14627 + }, + { + "path": "70.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14628 + }, + { + "path": "122.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "train", + "choice_d": "cafe/restaurant", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14629 + }, + { + "path": "229.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "bus", + "choice_b": "home", + "choice_c": "office", + "choice_d": "cafe/restaurant", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14630 + }, + { + "path": "1296.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14631 + }, + { + "path": "862.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14632 + }, + { + "path": "788.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14633 + }, + { + "path": "1082.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "city center", + "choice_d": "forest path", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14634 + }, + { + "path": "1491.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "tram", + "choice_b": "park", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14635 + }, + { + "path": "451.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14636 + }, + { + "path": "288.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "cafe/restaurant", + "choice_b": "home", + "choice_c": "grocery store", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14637 + }, + { + "path": "1063.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "cafe/restaurant", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14638 + }, + { + "path": "458.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "library", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "city center", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14639 + }, + { + "path": "816.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "library", + "choice_d": "forest path", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14640 + }, + { + "path": "1298.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14641 + }, + { + "path": "92.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "cafe/restaurant", + "choice_b": "beach", + "choice_c": "tram", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14642 + }, + { + "path": "127.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "car", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "tram", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14643 + }, + { + "path": "295.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "beach", + "choice_b": "office", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14644 + }, + { + "path": "290.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "forest path", + "choice_b": "train", + "choice_c": "grocery store", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14645 + }, + { + "path": "1148.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "home", + "choice_b": "forest path", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14646 + }, + { + "path": "720.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14647 + }, + { + "path": "688.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "residential area", + "choice_b": "home", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14648 + }, + { + "path": "784.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "office", + "choice_d": "forest path", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14649 + }, + { + "path": "182.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "library", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "metro station", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14650 + }, + { + "path": "820.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "park", + "choice_b": "home", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14651 + }, + { + "path": "101.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "bus", + "choice_b": "car", + "choice_c": "cafe/restaurant", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14652 + }, + { + "path": "1227.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "metro station", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "office", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14653 + }, + { + "path": "1436.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "library", + "choice_b": "residential area", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14654 + }, + { + "path": "384.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "tram", + "choice_b": "beach", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14655 + }, + { + "path": "234.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "residential area", + "choice_b": "office", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14656 + }, + { + "path": "102.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "forest path", + "choice_b": "tram", + "choice_c": "home", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14657 + }, + { + "path": "159.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14658 + }, + { + "path": "735.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "metro station", + "choice_b": "park", + "choice_c": "train", + "choice_d": "tram", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14659 + }, + { + "path": "401.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "car", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "city center", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14660 + }, + { + "path": "443.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14661 + }, + { + "path": "905.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "car", + "choice_b": "train", + "choice_c": "forest path", + "choice_d": "residential area", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14662 + }, + { + "path": "21.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "bus", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14663 + }, + { + "path": "1217.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "office", + "choice_d": "forest path", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14664 + }, + { + "path": "1147.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14665 + }, + { + "path": "764.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "city center", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14666 + }, + { + "path": "243.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14667 + }, + { + "path": "653.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "library", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14668 + }, + { + "path": "1061.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "tram", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14669 + }, + { + "path": "552.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "beach", + "choice_d": "city center", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14670 + }, + { + "path": "16.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "forest path", + "choice_b": "library", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14671 + }, + { + "path": "1322.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "train", + "choice_d": "residential area", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14672 + }, + { + "path": "1184.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "metro station", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14673 + }, + { + "path": "1215.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14674 + }, + { + "path": "971.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "cafe/restaurant", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14675 + }, + { + "path": "834.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "forest path", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14676 + }, + { + "path": "499.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "train", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14677 + }, + { + "path": "1292.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "metro station", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14678 + }, + { + "path": "236.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "car", + "choice_b": "train", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14679 + }, + { + "path": "793.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "office", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14680 + }, + { + "path": "1461.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "office", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14681 + }, + { + "path": "209.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "train", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14682 + }, + { + "path": "1039.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "city center", + "choice_b": "home", + "choice_c": "cafe/restaurant", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14683 + }, + { + "path": "29.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "library", + "choice_b": "car", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14684 + }, + { + "path": "860.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "grocery store", + "choice_b": "tram", + "choice_c": "office", + "choice_d": "home", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14685 + }, + { + "path": "35.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "park", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "home", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14686 + }, + { + "path": "663.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "residential area", + "choice_b": "office", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14687 + }, + { + "path": "60.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14688 + }, + { + "path": "676.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "office", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14689 + }, + { + "path": "285.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14690 + }, + { + "path": "1158.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "residential area", + "choice_b": "train", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14691 + }, + { + "path": "1616.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "train", + "choice_b": "home", + "choice_c": "park", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14692 + }, + { + "path": "1112.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "beach", + "choice_b": "metro station", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14693 + }, + { + "path": "512.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "office", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14694 + }, + { + "path": "675.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "forest path", + "choice_b": "city center", + "choice_c": "tram", + "choice_d": "home", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14695 + }, + { + "path": "496.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "residential area", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14696 + }, + { + "path": "1024.wav", + "question": "What place do the acoustic elements in this sound recording belong to?", + "choice_a": "cafe/restaurant", + "choice_b": "grocery store", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14697 + }, + { + "path": "1509.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14698 + }, + { + "path": "970.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "forest path", + "choice_d": "cafe/restaurant", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14699 + }, + { + "path": "847.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "train", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14700 + }, + { + "path": "760.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "forest path", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14701 + }, + { + "path": "1107.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "home", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14702 + }, + { + "path": "840.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "cafe/restaurant", + "choice_d": "metro station", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14703 + }, + { + "path": "1563.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "city center", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14704 + }, + { + "path": "305.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "forest path", + "choice_b": "metro station", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14705 + }, + { + "path": "601.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "library", + "choice_d": "tram", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14706 + }, + { + "path": "1611.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "bus", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14707 + }, + { + "path": "1121.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "home", + "choice_b": "bus", + "choice_c": "library", + "choice_d": "cafe/restaurant", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14708 + }, + { + "path": "154.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "city center", + "choice_b": "home", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14709 + }, + { + "path": "437.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "tram", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14710 + }, + { + "path": "766.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "tram", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14711 + }, + { + "path": "1380.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "tram", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14712 + }, + { + "path": "634.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14713 + }, + { + "path": "602.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "beach", + "choice_b": "bus", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14714 + }, + { + "path": "161.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14715 + }, + { + "path": "957.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "home", + "choice_b": "cafe/restaurant", + "choice_c": "library", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14716 + }, + { + "path": "1609.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "office", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "beach", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14717 + }, + { + "path": "62.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "home", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14718 + }, + { + "path": "613.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "library", + "choice_d": "residential area", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14719 + }, + { + "path": "1455.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "train", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "office", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14720 + }, + { + "path": "620.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "city center", + "choice_b": "beach", + "choice_c": "home", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14721 + }, + { + "path": "1398.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14722 + }, + { + "path": "366.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "bus", + "choice_b": "library", + "choice_c": "park", + "choice_d": "forest path", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14723 + }, + { + "path": "448.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "office", + "choice_b": "train", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14724 + }, + { + "path": "580.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "office", + "choice_b": "home", + "choice_c": "city center", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14725 + }, + { + "path": "516.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "tram", + "choice_b": "car", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14726 + }, + { + "path": "1178.wav", + "question": "Which of these environments do you associate with the recorded sounds?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14727 + }, + { + "path": "553.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14728 + }, + { + "path": "739.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "home", + "choice_b": "library", + "choice_c": "tram", + "choice_d": "forest path", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14729 + }, + { + "path": "538.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "train", + "choice_b": "grocery store", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14730 + }, + { + "path": "454.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "cafe/restaurant", + "choice_b": "office", + "choice_c": "train", + "choice_d": "tram", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14731 + }, + { + "path": "1599.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "bus", + "choice_b": "city center", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14732 + }, + { + "path": "1590.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "library", + "choice_b": "home", + "choice_c": "park", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14733 + }, + { + "path": "1308.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "residential area", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14734 + }, + { + "path": "106.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "park", + "choice_b": "residential area", + "choice_c": "bus", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14735 + }, + { + "path": "253.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "grocery store", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14736 + }, + { + "path": "649.wav", + "question": "Where might these sounds be commonly encountered?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "car", + "choice_d": "metro station", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14737 + }, + { + "path": "393.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "home", + "choice_b": "park", + "choice_c": "grocery store", + "choice_d": "library", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14738 + }, + { + "path": "647.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "beach", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14739 + }, + { + "path": "421.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "tram", + "choice_b": "grocery store", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14740 + }, + { + "path": "45.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "park", + "choice_d": "forest path", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14741 + }, + { + "path": "1268.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "library", + "choice_b": "city center", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14742 + }, + { + "path": "282.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14743 + }, + { + "path": "1131.wav", + "question": "Where does the nature of these sounds transport you?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "bus", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14744 + }, + { + "path": "1189.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "residential area", + "choice_b": "library", + "choice_c": "grocery store", + "choice_d": "home", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14745 + }, + { + "path": "1444.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "home", + "choice_b": "grocery store", + "choice_c": "tram", + "choice_d": "park", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14746 + }, + { + "path": "1058.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "train", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "tram", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14747 + }, + { + "path": "627.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "tram", + "choice_b": "city center", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14748 + }, + { + "path": "429.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "library", + "choice_b": "city center", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14749 + }, + { + "path": "989.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14750 + }, + { + "path": "961.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "library", + "choice_d": "grocery store", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14751 + }, + { + "path": "358.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "home", + "choice_b": "library", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14752 + }, + { + "path": "1160.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "library", + "choice_d": "car", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14753 + }, + { + "path": "1418.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14754 + }, + { + "path": "1388.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "cafe/restaurant", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14755 + }, + { + "path": "988.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "beach", + "choice_b": "residential area", + "choice_c": "tram", + "choice_d": "library", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14756 + }, + { + "path": "1051.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "grocery store", + "choice_b": "cafe/restaurant", + "choice_c": "office", + "choice_d": "beach", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14757 + }, + { + "path": "323.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "tram", + "choice_b": "bus", + "choice_c": "forest path", + "choice_d": "grocery store", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14758 + }, + { + "path": "1527.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "office", + "choice_d": "tram", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14759 + }, + { + "path": "1498.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "grocery store", + "choice_b": "train", + "choice_c": "residential area", + "choice_d": "beach", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14760 + }, + { + "path": "693.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "city center", + "choice_b": "library", + "choice_c": "car", + "choice_d": "grocery store", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14761 + }, + { + "path": "1603.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "park", + "choice_b": "library", + "choice_c": "train", + "choice_d": "beach", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14762 + }, + { + "path": "794.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "home", + "choice_b": "car", + "choice_c": "park", + "choice_d": "city center", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14763 + }, + { + "path": "1238.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "bus", + "choice_b": "forest path", + "choice_c": "park", + "choice_d": "office", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14764 + }, + { + "path": "1480.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14765 + }, + { + "path": "1019.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14766 + }, + { + "path": "903.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14767 + }, + { + "path": "1432.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "metro station", + "choice_b": "tram", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14768 + }, + { + "path": "645.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "grocery store", + "choice_b": "cafe/restaurant", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14769 + }, + { + "path": "962.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "residential area", + "choice_b": "beach", + "choice_c": "cafe/restaurant", + "choice_d": "forest path", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14770 + }, + { + "path": "1425.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "city center", + "choice_b": "cafe/restaurant", + "choice_c": "car", + "choice_d": "forest path", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14771 + }, + { + "path": "902.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "residential area", + "choice_d": "park", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14772 + }, + { + "path": "837.wav", + "question": "What is the likely source of these sounds?", + "choice_a": "forest path", + "choice_b": "city center", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14773 + }, + { + "path": "547.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "beach", + "choice_b": "home", + "choice_c": "park", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14774 + }, + { + "path": "238.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "bus", + "choice_d": "cafe/restaurant", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14775 + }, + { + "path": "1135.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "forest path", + "choice_b": "grocery store", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14776 + }, + { + "path": "955.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "train", + "choice_b": "beach", + "choice_c": "park", + "choice_d": "residential area", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14777 + }, + { + "path": "203.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "train", + "choice_b": "beach", + "choice_c": "car", + "choice_d": "library", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14778 + }, + { + "path": "515.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "library", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "home", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14779 + }, + { + "path": "139.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "bus", + "choice_b": "office", + "choice_c": "metro station", + "choice_d": "home", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14780 + }, + { + "path": "916.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "metro station", + "choice_b": "office", + "choice_c": "train", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14781 + }, + { + "path": "1162.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "home", + "choice_b": "residential area", + "choice_c": "car", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14782 + }, + { + "path": "1077.wav", + "question": "What is the primary setting indicated by the sounds in the recording?", + "choice_a": "train", + "choice_b": "metro station", + "choice_c": "park", + "choice_d": "library", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14783 + }, + { + "path": "1487.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "bus", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14784 + }, + { + "path": "372.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "home", + "choice_b": "park", + "choice_c": "cafe/restaurant", + "choice_d": "train", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14785 + }, + { + "path": "185.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "forest path", + "choice_b": "bus", + "choice_c": "metro station", + "choice_d": "residential area", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14786 + }, + { + "path": "911.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "park", + "choice_b": "bus", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14787 + }, + { + "path": "628.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "library", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14788 + }, + { + "path": "1016.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "grocery store", + "choice_b": "car", + "choice_c": "office", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14789 + }, + { + "path": "255.wav", + "question": "From the sound, where do you picture this recording was made?", + "choice_a": "forest path", + "choice_b": "cafe/restaurant", + "choice_c": "bus", + "choice_d": "beach", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14790 + }, + { + "path": "125.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "city center", + "choice_b": "office", + "choice_c": "beach", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14791 + }, + { + "path": "1562.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "cafe/restaurant", + "choice_d": "park", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14792 + }, + { + "path": "670.wav", + "question": "Which location seems to match the sonic landscape?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14793 + }, + { + "path": "464.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "home", + "choice_b": "forest path", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14794 + }, + { + "path": "796.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "home", + "choice_b": "train", + "choice_c": "office", + "choice_d": "bus", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14795 + }, + { + "path": "1477.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "forest path", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "cafe/restaurant", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14796 + }, + { + "path": "899.wav", + "question": "What scenario can you infer from the auditory cues in this sound clip?", + "choice_a": "residential area", + "choice_b": "city center", + "choice_c": "library", + "choice_d": "bus", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14797 + }, + { + "path": "1010.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "train", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14798 + }, + { + "path": "1337.wav", + "question": "What backdrop do the sounds in the recording paint for you?", + "choice_a": "cafe/restaurant", + "choice_b": "train", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14799 + }, + { + "path": "103.wav", + "question": "Where do you believe this audio clip was captured?", + "choice_a": "grocery store", + "choice_b": "city center", + "choice_c": "library", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14800 + }, + { + "path": "491.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "tram", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14801 + }, + { + "path": "279.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "train", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "office", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14802 + }, + { + "path": "700.wav", + "question": "What surroundings are suggested by the audible background?", + "choice_a": "forest path", + "choice_b": "office", + "choice_c": "residential area", + "choice_d": "home", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14803 + }, + { + "path": "1378.wav", + "question": "In what context do these sounds typically occur?", + "choice_a": "library", + "choice_b": "office", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14804 + }, + { + "path": "276.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "grocery store", + "choice_b": "home", + "choice_c": "metro station", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14805 + }, + { + "path": "898.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "beach", + "choice_b": "library", + "choice_c": "train", + "choice_d": "city center", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14806 + }, + { + "path": "272.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "beach", + "choice_b": "car", + "choice_c": "metro station", + "choice_d": "park", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14807 + }, + { + "path": "422.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "car", + "choice_b": "forest path", + "choice_c": "cafe/restaurant", + "choice_d": "library", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14808 + }, + { + "path": "950.wav", + "question": "What is the most fitting scene for the sounds you're hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "park", + "choice_c": "beach", + "choice_d": "residential area", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14809 + }, + { + "path": "1304.wav", + "question": "What is the setting that comes to mind when listening to this sound?", + "choice_a": "city center", + "choice_b": "residential area", + "choice_c": "beach", + "choice_d": "cafe/restaurant", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14810 + }, + { + "path": "1086.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "train", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "residential area", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14811 + }, + { + "path": "1065.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "cafe/restaurant", + "choice_b": "metro station", + "choice_c": "home", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14812 + }, + { + "path": "912.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "park", + "choice_d": "home", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14813 + }, + { + "path": "359.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "park", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "metro station", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14814 + }, + { + "path": "1588.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "cafe/restaurant", + "choice_b": "forest path", + "choice_c": "office", + "choice_d": "city center", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14815 + }, + { + "path": "1098.wav", + "question": "What location does the sound atmosphere point to?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "grocery store", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14816 + }, + { + "path": "1021.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "train", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14817 + }, + { + "path": "432.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "home", + "choice_b": "tram", + "choice_c": "grocery store", + "choice_d": "bus", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14818 + }, + { + "path": "1232.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "car", + "choice_b": "cafe/restaurant", + "choice_c": "forest path", + "choice_d": "beach", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14819 + }, + { + "path": "1428.wav", + "question": "What setting corresponds to the audio cues present?", + "choice_a": "home", + "choice_b": "forest path", + "choice_c": "office", + "choice_d": "metro station", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14820 + }, + { + "path": "1438.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "grocery store", + "choice_b": "tram", + "choice_c": "cafe/restaurant", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14821 + }, + { + "path": "189.wav", + "question": "What type of location do the sounds in the recording hint at?", + "choice_a": "city center", + "choice_b": "forest path", + "choice_c": "home", + "choice_d": "library", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14822 + }, + { + "path": "237.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "forest path", + "choice_d": "cafe/restaurant", + "answer_gt": "library", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14823 + }, + { + "path": "548.wav", + "question": "Where do the sounds in the audio capture seem most at home?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "beach", + "choice_d": "home", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14824 + }, + { + "path": "1415.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "bus", + "choice_d": "park", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14825 + }, + { + "path": "3.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "beach", + "choice_b": "cafe/restaurant", + "choice_c": "metro station", + "choice_d": "forest path", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14826 + }, + { + "path": "1168.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14827 + }, + { + "path": "978.wav", + "question": "Where do the unique sound signatures lead you to believe the recording took place?", + "choice_a": "office", + "choice_b": "residential area", + "choice_c": "cafe/restaurant", + "choice_d": "tram", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14828 + }, + { + "path": "728.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "metro station", + "choice_b": "bus", + "choice_c": "home", + "choice_d": "residential area", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14829 + }, + { + "path": "832.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "tram", + "choice_b": "residential area", + "choice_c": "office", + "choice_d": "forest path", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14830 + }, + { + "path": "136.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "residential area", + "choice_b": "park", + "choice_c": "car", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14831 + }, + { + "path": "598.wav", + "question": "What scenery does the ambient sound reflect?", + "choice_a": "park", + "choice_b": "city center", + "choice_c": "residential area", + "choice_d": "tram", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14832 + }, + { + "path": "1530.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "cafe/restaurant", + "choice_b": "park", + "choice_c": "tram", + "choice_d": "train", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14833 + }, + { + "path": "104.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "forest path", + "choice_b": "beach", + "choice_c": "city center", + "choice_d": "park", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14834 + }, + { + "path": "378.wav", + "question": "Which location do these sounds seem to illustrate?", + "choice_a": "residential area", + "choice_b": "home", + "choice_c": "forest path", + "choice_d": "grocery store", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14835 + }, + { + "path": "716.wav", + "question": "Where do the specific sounds heard in the recording typically occur?", + "choice_a": "metro station", + "choice_b": "car", + "choice_c": "train", + "choice_d": "grocery store", + "answer_gt": "grocery store", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14836 + }, + { + "path": "354.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "home", + "choice_b": "cafe/restaurant", + "choice_c": "park", + "choice_d": "grocery store", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14837 + }, + { + "path": "1287.wav", + "question": "What location comes to mind with this collection of sounds?", + "choice_a": "tram", + "choice_b": "train", + "choice_c": "city center", + "choice_d": "metro station", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14838 + }, + { + "path": "1320.wav", + "question": "What does the sonic environment in the recording suggest?", + "choice_a": "cafe/restaurant", + "choice_b": "bus", + "choice_c": "beach", + "choice_d": "forest path", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14839 + }, + { + "path": "337.wav", + "question": "What location does the background noise suggest to you?", + "choice_a": "grocery store", + "choice_b": "cafe/restaurant", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14840 + }, + { + "path": "436.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "grocery store", + "choice_b": "metro station", + "choice_c": "car", + "choice_d": "bus", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14841 + }, + { + "path": "1593.wav", + "question": "In what kind of area do these types of sounds usually arise?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "city center", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14842 + }, + { + "path": "1419.wav", + "question": "In which of these places would you most expect to hear these sounds?", + "choice_a": "beach", + "choice_b": "metro station", + "choice_c": "train", + "choice_d": "library", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14843 + }, + { + "path": "351.wav", + "question": "Which of these options aligns with the sound characteristics you hear?", + "choice_a": "city center", + "choice_b": "grocery store", + "choice_c": "home", + "choice_d": "beach", + "answer_gt": "home", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14844 + }, + { + "path": "1523.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "residential area", + "choice_b": "bus", + "choice_c": "car", + "choice_d": "city center", + "answer_gt": "residential area", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14845 + }, + { + "path": "41.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "car", + "choice_b": "bus", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14846 + }, + { + "path": "974.wav", + "question": "What scene does the layered soundscape suggest to you?", + "choice_a": "beach", + "choice_b": "city center", + "choice_c": "cafe/restaurant", + "choice_d": "residential area", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14847 + }, + { + "path": "1169.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "metro station", + "choice_b": "beach", + "choice_c": "train", + "choice_d": "library", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14848 + }, + { + "path": "687.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "car", + "choice_b": "cafe/restaurant", + "choice_c": "train", + "choice_d": "home", + "answer_gt": "car", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14849 + }, + { + "path": "972.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "office", + "choice_b": "grocery store", + "choice_c": "residential area", + "choice_d": "home", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14850 + }, + { + "path": "17.wav", + "question": "What setting are you transported to with these sounds?", + "choice_a": "bus", + "choice_b": "grocery store", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14851 + }, + { + "path": "1313.wav", + "question": "Which place matches the ambient sounds you are hearing?", + "choice_a": "tram", + "choice_b": "office", + "choice_c": "train", + "choice_d": "forest path", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14852 + }, + { + "path": "1305.wav", + "question": "What place do you connect with the sound profile in this clip?", + "choice_a": "tram", + "choice_b": "cafe/restaurant", + "choice_c": "office", + "choice_d": "residential area", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14853 + }, + { + "path": "483.wav", + "question": "What setting would you deduce from the sound characteristics?", + "choice_a": "library", + "choice_b": "residential area", + "choice_c": "forest path", + "choice_d": "city center", + "answer_gt": "city center", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14854 + }, + { + "path": "153.wav", + "question": "What scene does the combination of sounds evoke?", + "choice_a": "metro station", + "choice_b": "forest path", + "choice_c": "office", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14855 + }, + { + "path": "1533.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14856 + }, + { + "path": "990.wav", + "question": "Where could these sound elements commonly be found?", + "choice_a": "beach", + "choice_b": "park", + "choice_c": "residential area", + "choice_d": "train", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14857 + }, + { + "path": "204.wav", + "question": "Where can you envision this sound taking place?", + "choice_a": "forest path", + "choice_b": "train", + "choice_c": "home", + "choice_d": "car", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14858 + }, + { + "path": "59.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "residential area", + "choice_b": "home", + "choice_c": "train", + "choice_d": "metro station", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14859 + }, + { + "path": "353.wav", + "question": "Where does the acoustic environment place you?", + "choice_a": "car", + "choice_b": "library", + "choice_c": "metro station", + "choice_d": "tram", + "answer_gt": "tram", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14860 + }, + { + "path": "1331.wav", + "question": "Which of these locations do the sounds give the impression of being recorded in?", + "choice_a": "metro station", + "choice_b": "cafe/restaurant", + "choice_c": "grocery store", + "choice_d": "city center", + "answer_gt": "metro station", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14861 + }, + { + "path": "293.wav", + "question": "What environment is being reflected in this sound recording?", + "choice_a": "cafe/restaurant", + "choice_b": "car", + "choice_c": "tram", + "choice_d": "bus", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14862 + }, + { + "path": "578.wav", + "question": "Which of the following locations best fits the soundscape?", + "choice_a": "home", + "choice_b": "office", + "choice_c": "cafe/restaurant", + "choice_d": "residential area", + "answer_gt": "cafe/restaurant", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14863 + }, + { + "path": "1502.wav", + "question": "In what setting do you think this sound was recorded?", + "choice_a": "forest path", + "choice_b": "library", + "choice_c": "bus", + "choice_d": "grocery store", + "answer_gt": "forest path", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14864 + }, + { + "path": "140.wav", + "question": "Where is the most probable place these sounds originate from?", + "choice_a": "beach", + "choice_b": "forest path", + "choice_c": "residential area", + "choice_d": "grocery store", + "answer_gt": "beach", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14865 + }, + { + "path": "195.wav", + "question": "What venue are the sounds indicative of?", + "choice_a": "metro station", + "choice_b": "home", + "choice_c": "library", + "choice_d": "office", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14866 + }, + { + "path": "206.wav", + "question": "Where do these distinct sounds come from?", + "choice_a": "office", + "choice_b": "bus", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14867 + }, + { + "path": "1291.wav", + "question": "What scene is being depicted through these sounds?", + "choice_a": "cafe/restaurant", + "choice_b": "home", + "choice_c": "library", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14868 + }, + { + "path": "1167.wav", + "question": "What venue do these sounds most likely represent?", + "choice_a": "home", + "choice_b": "car", + "choice_c": "library", + "choice_d": "park", + "answer_gt": "park", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14869 + }, + { + "path": "1015.wav", + "question": "What is your best guess at the setting of this sound clip?", + "choice_a": "bus", + "choice_b": "city center", + "choice_c": "cafe/restaurant", + "choice_d": "office", + "answer_gt": "bus", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14870 + }, + { + "path": "1235.wav", + "question": "Where do the audible clues in this recording lead you to believe it was taken?", + "choice_a": "park", + "choice_b": "office", + "choice_c": "forest path", + "choice_d": "train", + "answer_gt": "office", + "task_name": "Acoustic_Scene_Classification", + "dataset_name": "TUT2017", + "uniq_id": 14871 + }, + { + "path": "Rain with thunder in a city.wav", + "question": "What weather phenomenon is heard?", + "choice_a": "fireworks", + "choice_b": "mouse", + "choice_c": "yes", + "choice_d": "thunder", + "answer_gt": "thunder", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14872 + }, + { + "path": "Gentle rain outside balcony street noise.wav", + "question": "Is there more sounds than just an engine?", + "choice_a": "seesaw", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14873 + }, + { + "path": "Running.wav", + "question": "does the person walk then run?", + "choice_a": "eating", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14874 + }, + { + "path": "April dawn chorus Sydenham Hill.wav", + "question": "How many birds are chirping?", + "choice_a": "birds", + "choice_b": "small", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14875 + }, + { + "path": "20080416.buzz.stereo.wav", + "question": "Is there an animal with fur nearby?", + "choice_a": "ground", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14876 + }, + { + "path": "Crunchy walk on pebbles.wav", + "question": "What is the person doing?", + "choice_a": "walking", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14877 + }, + { + "path": "Idling Truck at Loading Dock.wav", + "question": "Are people speaking?", + "choice_a": "sound", + "choice_b": "party", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14878 + }, + { + "path": "20130326_caged.birds.01.wav", + "question": "Are the birds chirping?", + "choice_a": "rain", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14879 + }, + { + "path": "tornado day 1.wav", + "question": "Is the machine heard starting up and shutting down?", + "choice_a": "one", + "choice_b": "flood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14880 + }, + { + "path": "maryam sounds 5.wav", + "question": "is the item solid?", + "choice_a": "running", + "choice_b": "bucket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14881 + }, + { + "path": "plastic-straw-whistles.wav", + "question": "Does it sound like a wind instrument?", + "choice_a": "cars", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14882 + }, + { + "path": "CAR_WASH.wav", + "question": "Is a vehicle moving in the distance?", + "choice_a": "six", + "choice_b": "snow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14883 + }, + { + "path": "fdv_orage_26082011.wav", + "question": "Is outside sunny?", + "choice_a": "traffic", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14884 + }, + { + "path": "AMB_swamp_summer_night_fish_insects_00.wav", + "question": "Are people dancing at a pool party?", + "choice_a": "music", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14885 + }, + { + "path": "WindInPylons.wav", + "question": "Is the wind blowing?", + "choice_a": "constant", + "choice_b": "never", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14886 + }, + { + "path": "Streatham Railway Station and on a Train.wav", + "question": "Can humans be heard talking?", + "choice_a": "short", + "choice_b": "rod", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14887 + }, + { + "path": "Bell_Hotel_desk.wav", + "question": "Does the bell stop and start?", + "choice_a": "thunder", + "choice_b": "interested", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14888 + }, + { + "path": "Padlock.wav", + "question": "What could be making this clicking and scratching noise?", + "choice_a": "lock", + "choice_b": "bike", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "lock", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14889 + }, + { + "path": "Outside wind.wav", + "question": "Is the chair squeaking?", + "choice_a": "honking", + "choice_b": "stormy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14890 + }, + { + "path": "Spring Birds Raw _New Jersey_.wav", + "question": "Is this something that can be heard on an airplane?", + "choice_a": "people", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14891 + }, + { + "path": "windy rain.wav", + "question": "What is making the constant sound?", + "choice_a": "screeching", + "choice_b": "siren", + "choice_c": "rain", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14892 + }, + { + "path": "heating_far away.wav", + "question": "Is a parrot squawking?", + "choice_a": "hall", + "choice_b": "honking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14893 + }, + { + "path": "California morning birds singing.wav", + "question": "Is there a storm coming in?", + "choice_a": "bell", + "choice_b": "mouse", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14894 + }, + { + "path": "rain_medium_thunders.wav", + "question": "Is it raining?", + "choice_a": "bird", + "choice_b": "never", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14895 + }, + { + "path": "breakfast ambience.wav", + "question": "how many people are speaking?", + "choice_a": "driller", + "choice_b": "zoo", + "choice_c": "two", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14896 + }, + { + "path": "Subway-Moscow-013.wav", + "question": "Is there a tractor in the background?", + "choice_a": "crickets", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14897 + }, + { + "path": "abandoned-ballroom-radiators.wav", + "question": "Is the wind blowing?", + "choice_a": "small", + "choice_b": "crows", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14898 + }, + { + "path": "Heavy Rain 1.wav", + "question": "Is the water falling onto metal?", + "choice_a": "men", + "choice_b": "twice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14899 + }, + { + "path": "london-st-james-park-feeding-the-birds.wav", + "question": "What type of animal would make these sounds?", + "choice_a": "birds", + "choice_b": "machine", + "choice_c": "seagulls", + "choice_d": "no", + "answer_gt": "seagulls", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14900 + }, + { + "path": "Mariehamn_frogs.wav", + "question": "how many quacks are there?", + "choice_a": "twentyfive", + "choice_b": "radio", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "twentyfive", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14901 + }, + { + "path": "drunk_teenagers_1.wav", + "question": "is one person screaming louder than others?", + "choice_a": "eight", + "choice_b": "forty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14902 + }, + { + "path": "Waiting at a Montreal Subway Station.wav", + "question": "Are animals screaming?", + "choice_a": "screeching", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14903 + }, + { + "path": "earth_movement.wav", + "question": "This sound is reminiscent of what domesticated animal when it's happy?", + "choice_a": "chain", + "choice_b": "nothing", + "choice_c": "cat", + "choice_d": "no", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14904 + }, + { + "path": "Metal_On_Wood_Hits_Axe.wav", + "question": "Is the sound interrupted at any point?", + "choice_a": "umbrella", + "choice_b": "seven", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14905 + }, + { + "path": "car-radio-am-noise2.wav", + "question": "Is the sound high pitched?", + "choice_a": "never", + "choice_b": "radio", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14906 + }, + { + "path": "Greek Habitues - _Evosmos - Salonika_ 16.18 28.09.wav", + "question": "Are there kids?", + "choice_a": "slowly", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14907 + }, + { + "path": "Machetes sliding 2.wav", + "question": "Is the person sometimes striking an object and occasionally sliding an object across another?", + "choice_a": "crickets", + "choice_b": "four", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14908 + }, + { + "path": "bebops_water1.wav", + "question": "is water making a noise?", + "choice_a": "falls", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14909 + }, + { + "path": "Cicadas .wav", + "question": "Is there a shaking sound?", + "choice_a": "electronic", + "choice_b": "seven", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14910 + }, + { + "path": "Creacking Oak 6bft SHORT 130418_00.wav", + "question": "What animal type is making the sound at the end?", + "choice_a": "birds", + "choice_b": "big", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14911 + }, + { + "path": "dripping.wav", + "question": "Is water babbling?", + "choice_a": "chirp", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14912 + }, + { + "path": "squeaky metal swing.wav", + "question": "What item is being moved back and forth to create the noise?", + "choice_a": "no", + "choice_b": "nine", + "choice_c": "cat", + "choice_d": "no", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14913 + }, + { + "path": "Autumnal Ambient 24 Bits 48 Khz.wav", + "question": "Are people splashing the water?", + "choice_a": "waves", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14914 + }, + { + "path": "Fryers Forest - Powerful Owl _Ninox Stenua_.wav", + "question": "is this likely in the daytime?", + "choice_a": "hands", + "choice_b": "hand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14915 + }, + { + "path": "metal plate striking wall.wav", + "question": "does the metal seem to be picked up and dropped again?", + "choice_a": "bear", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14916 + }, + { + "path": "Beer Pong Sounds ball table and cups.wav", + "question": "What falls before the tapping begins?", + "choice_a": "ocean", + "choice_b": "four", + "choice_c": "ball", + "choice_d": "no", + "answer_gt": "ball", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14917 + }, + { + "path": "Canada Geese Squawk on a Pond with a Fountain.wav", + "question": "Is this in the city?", + "choice_a": "lock", + "choice_b": "men", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14918 + }, + { + "path": "FM Radio Tuning Sweep.wav", + "question": "Is there a lot of static?", + "choice_a": "coin", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14919 + }, + { + "path": "Toilet Shuffling.wav", + "question": "What is the water doing?", + "choice_a": "intercom", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "flushing", + "answer_gt": "flushing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14920 + }, + { + "path": "rain_medium_thunders.wav", + "question": "Is this noise associated with dryness?", + "choice_a": "stormy", + "choice_b": "clapping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14921 + }, + { + "path": "RadioFan.wav", + "question": "What is being played in this clip?", + "choice_a": "music", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "music", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14922 + }, + { + "path": "Silencyo_CC_Wind Mistral_Medium_Dry Grass_Crickets_Distant traffic.wav", + "question": "Is someone walking loudly?", + "choice_a": "wet", + "choice_b": "raining", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14923 + }, + { + "path": "Wind-up Toy Motorbike SFX.wav", + "question": "Is a man sneezing?", + "choice_a": "window", + "choice_b": "short", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14924 + }, + { + "path": "water slushing moderate speed.wav", + "question": "Where is this liquid located?", + "choice_a": "bucket", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bucket", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14925 + }, + { + "path": "Rushing_water+wind-Rec_Samsung_HMX-F80_Camcorder.wav", + "question": "Is it wet around here?", + "choice_a": "four", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14926 + }, + { + "path": "Beach Wave Ambince .wav", + "question": "Is this person near the beach?", + "choice_a": "ten", + "choice_b": "clanging", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14927 + }, + { + "path": "Paper_Parchment_Rustling.wav", + "question": "Is the woman swimming in the pool?", + "choice_a": "stops", + "choice_b": "six", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14928 + }, + { + "path": "SCC CLAPTER 20101210.wav", + "question": "Is there only a single person in the audience?", + "choice_a": "falling", + "choice_b": "tearing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14929 + }, + { + "path": "Walking across carpeted floor with slippers.wav", + "question": "Does someone talk?", + "choice_a": "windy", + "choice_b": "buzzing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14930 + }, + { + "path": "Fence Hit_City ambience night.wav", + "question": "Is the sound made by a machine?", + "choice_a": "bus", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14931 + }, + { + "path": "wind-sound-from-inside-car.wav", + "question": "How many voices can be heard yelling in the inclement weather?", + "choice_a": "zero", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "zero", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14932 + }, + { + "path": "PIT-ROIG 0.12-0.17.wav", + "question": "What is the volume of this?", + "choice_a": "sand", + "choice_b": "tap", + "choice_c": "yes", + "choice_d": "normal", + "answer_gt": "normal", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14933 + }, + { + "path": "Padlock.wav", + "question": "Is this the sound of wood?", + "choice_a": "voice", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14934 + }, + { + "path": "incheon_4am.wav", + "question": "can voices be heard?", + "choice_a": "wrench", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14935 + }, + { + "path": "MISC_Int_Cat_Purring_002.wav", + "question": "Is the cat in a good mood?", + "choice_a": "fireworks", + "choice_b": "zoo", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14936 + }, + { + "path": "abandoned-ballroom-big-metal.wav", + "question": "Does anybody do any walking?", + "choice_a": "rooster", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14937 + }, + { + "path": "Walking alongside the road.wav", + "question": "Does a vehicle drive by?", + "choice_a": "cow", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14938 + }, + { + "path": "water fountian-hollow sm med pitch.wav", + "question": "Is the person singing in the shower?", + "choice_a": "motor", + "choice_b": "cars", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14939 + }, + { + "path": "20070824.supper.wav", + "question": "Is this outside?", + "choice_a": "heavy", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14940 + }, + { + "path": "Garbage Truck.wav", + "question": "Is machinery being used?", + "choice_a": "three", + "choice_b": "electronic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14941 + }, + { + "path": "Construction Zone.wav", + "question": "Is the vehicle getting closer?", + "choice_a": "keys", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14942 + }, + { + "path": "AC Unit.wav", + "question": "Are there animals around?", + "choice_a": "wind", + "choice_b": "eating", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14943 + }, + { + "path": "Sizzling Bacon.wav", + "question": "Is there liquid?", + "choice_a": "music", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14944 + }, + { + "path": "Toilet Shuffling.wav", + "question": "What object was just used?", + "choice_a": "singing", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "closet", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14945 + }, + { + "path": "porto_morning_tropical_birds_market_20.wav", + "question": "Are there any birds?", + "choice_a": "birds", + "choice_b": "fifteen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14946 + }, + { + "path": "Cardiff Bay fireworks.wav", + "question": "What is making the load bangs?", + "choice_a": "road", + "choice_b": "chain", + "choice_c": "yes", + "choice_d": "fireworks", + "answer_gt": "fireworks", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14947 + }, + { + "path": "Waiting at a Montreal Subway Station.wav", + "question": "Is it played by musical instruments?", + "choice_a": "lock", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14948 + }, + { + "path": "Ext, distance village, light wind in tree-01.wav", + "question": "Is a cow mooing?", + "choice_a": "traffic", + "choice_b": "announcement", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14949 + }, + { + "path": "London Overground train _interior_ approaches Victoria Station.wav", + "question": "What are the people on the train doing?", + "choice_a": "low", + "choice_b": "intercom", + "choice_c": "yes", + "choice_d": "talking", + "answer_gt": "talking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14950 + }, + { + "path": "fountains-Udaipur-Saheliyon-Ki-Bari-4.wav", + "question": "Is the ground wet?", + "choice_a": "running", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14951 + }, + { + "path": "Basement Water Pump.wav", + "question": "Does the motor run smoothly?", + "choice_a": "ocean", + "choice_b": "screeching", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14952 + }, + { + "path": "tua-mirandela_train_arrival_march2007.wav", + "question": "Is the person talking?", + "choice_a": "glass", + "choice_b": "fireworks", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14953 + }, + { + "path": "FOOTSTEPS_005.wav", + "question": "What is the person walking on?", + "choice_a": "crickets", + "choice_b": "snow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "snow", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14954 + }, + { + "path": "public.wav", + "question": "Is this a quiet place?", + "choice_a": "umbrella", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14955 + }, + { + "path": "20101228.teens.wav", + "question": "What noise does the group make with their hands?", + "choice_a": "nothing", + "choice_b": "whispering", + "choice_c": "clapping", + "choice_d": "no", + "answer_gt": "clapping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14956 + }, + { + "path": "Alps village field-recording distance.wav", + "question": "Are they dogs barking in the background?", + "choice_a": "one", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14957 + }, + { + "path": "Vomit, puking spilling water onto grass splat.wav", + "question": "Is something being poured out?", + "choice_a": "insect", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14958 + }, + { + "path": "Stream # 2.wav", + "question": "Is the sound of the falling rain?", + "choice_a": "normal", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14959 + }, + { + "path": "Car vs. Freight Train.wav", + "question": "Can birds be heard chirping?", + "choice_a": "sparrow", + "choice_b": "thunder", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14960 + }, + { + "path": "Slushing in mouth.wav", + "question": "Is the person holding something?", + "choice_a": "honking", + "choice_b": "short", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14961 + }, + { + "path": "1400 am static.wav", + "question": "Are there dogs in the background?", + "choice_a": "lock", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14962 + }, + { + "path": "Ronda - Fountain near the Town Hall _general_ - Fuente cerca del Ayuntamiento _general_.wav", + "question": "What object is sometimes carried by a person to stay dry when this is happening?", + "choice_a": "sheep", + "choice_b": "ocean", + "choice_c": "umbrella", + "choice_d": "no", + "answer_gt": "umbrella", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14963 + }, + { + "path": "slam.wav", + "question": "Is that sound of something being hit?", + "choice_a": "driving", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14964 + }, + { + "path": "Gentle rain outside balcony street noise.wav", + "question": "Does the engine stop making sound?", + "choice_a": "running", + "choice_b": "chirp", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14965 + }, + { + "path": "Munich Cathedral.wav", + "question": "Can people be heard talking?", + "choice_a": "coin", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14966 + }, + { + "path": "HammerDrill.wav", + "question": "Is a child giggling?", + "choice_a": "announcement", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14967 + }, + { + "path": "20091211.barking.stairs.wav", + "question": "What animal is making loud noise?", + "choice_a": "dog", + "choice_b": "grass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14968 + }, + { + "path": "RG Large Old Dog Snoring.wav", + "question": "Is that a car honking?", + "choice_a": "vehicle", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14969 + }, + { + "path": "080902_00_machine_generators.wav", + "question": "Are there several animals making noise?", + "choice_a": "running", + "choice_b": "yes", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14970 + }, + { + "path": "SilverStarSearchAndRescue.wav", + "question": "Does the buzzing get louder before dying off?", + "choice_a": "tree", + "choice_b": "ground", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14971 + }, + { + "path": "Krankenwagen _ German Ambulances Passing by...wav", + "question": "Is the siren getting louder?", + "choice_a": "speaking", + "choice_b": "train", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14972 + }, + { + "path": "windup_dino_slow.wav", + "question": "Is there a click clack sound throughout?", + "choice_a": "women", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14973 + }, + { + "path": "20110422_shower.wav", + "question": "Is it a sunny clear day?", + "choice_a": "dog", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14974 + }, + { + "path": "Cafeteria Ambience.wav", + "question": "Is this outdoors?", + "choice_a": "rod", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14975 + }, + { + "path": "Thunder Outside.wav", + "question": "Are there three strikes of thunder?", + "choice_a": "forty", + "choice_b": "ground", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14976 + }, + { + "path": "bridge demolition pounding.wav", + "question": "How many impacts can we hear ?", + "choice_a": "twelve", + "choice_b": "twenty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "twenty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14977 + }, + { + "path": "20110220_churchbell.wav", + "question": "How many times is the gong struck?", + "choice_a": "tearing", + "choice_b": "twenty", + "choice_c": "yes", + "choice_d": "four", + "answer_gt": "four", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14978 + }, + { + "path": "Jet Engine 1.wav", + "question": "What product is being processed?", + "choice_a": "fifteen", + "choice_b": "metal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "metal", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14979 + }, + { + "path": "Building Site.wav", + "question": "Is there more than one source of noise?", + "choice_a": "chirping", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14980 + }, + { + "path": "Jet Engine 1.wav", + "question": "is the saw changing in rhythm?", + "choice_a": "faucet", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14981 + }, + { + "path": "INT Factory budimka, pozega.wav", + "question": "Does the flat slapping at the beginning happen more than once?", + "choice_a": "car", + "choice_b": "motor", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14982 + }, + { + "path": "radio tuning 2.wav", + "question": "Can this device play music?", + "choice_a": "stormy", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14983 + }, + { + "path": "Subping03.wav", + "question": "Is there only a single sound?", + "choice_a": "train", + "choice_b": "drill", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14984 + }, + { + "path": "T24_home_dishwasher_cycle change.wav", + "question": "What was passing nearby?", + "choice_a": "lock", + "choice_b": "three", + "choice_c": "nothing", + "choice_d": "no", + "answer_gt": "nothing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14985 + }, + { + "path": "Underwater_rumble_CsG.wav", + "question": "Does the rumbling change tone at all?", + "choice_a": "cars", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14986 + }, + { + "path": "bellaromani.wav", + "question": "Is a guitar being played?", + "choice_a": "forty", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14987 + }, + { + "path": "winter wren wind leaves.wav", + "question": "Is the bird chirping loudly?", + "choice_a": "raining", + "choice_b": "whispering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14988 + }, + { + "path": "WaterBottle.wav", + "question": "Is the person making music?", + "choice_a": "ground", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14989 + }, + { + "path": "windup_dino_slow.wav", + "question": "Does it sound like something found in nature?", + "choice_a": "short", + "choice_b": "no", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14990 + }, + { + "path": "020220_00.wav", + "question": "Does the man take pauses as he speaks?", + "choice_a": "two", + "choice_b": "eight", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14991 + }, + { + "path": "Footsteps on Wet Pavement_1-2.wav", + "question": "Is there jumping?", + "choice_a": "slowly", + "choice_b": "swimming", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14992 + }, + { + "path": "glas-bubbels-def01.wav", + "question": "Is someone drinking something?", + "choice_a": "bird", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14993 + }, + { + "path": "Galle Road.wav", + "question": "What makes the squeaking sound at the start of the clip?", + "choice_a": "twice", + "choice_b": "bear", + "choice_c": "bus", + "choice_d": "no", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14994 + }, + { + "path": "Laser printer.wav", + "question": "Did the machine get unplugged?", + "choice_a": "chopping", + "choice_b": "loudspeaker", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14995 + }, + { + "path": "Bicin_Diputacion_Day_22-03-2009.wav", + "question": "does the clicking increase in speed?", + "choice_a": "rooster", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14996 + }, + { + "path": "Lincoln Nebraska Tornado 5 9 2016.wav", + "question": "How many times does the siren get louder?", + "choice_a": "band", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14997 + }, + { + "path": "pencil on paper.wav", + "question": "Are they writing slowly?", + "choice_a": "sheep", + "choice_b": "car", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14998 + }, + { + "path": "Balloon Game at Arlington Heights Carnival.wav", + "question": "Do they have bad manners?", + "choice_a": "clanging", + "choice_b": "motor", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 14999 + }, + { + "path": "OrchestraTuning2.wav", + "question": "Are multiple different instruments being played?", + "choice_a": "band", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15000 + }, + { + "path": "STE-031 italian road.wav", + "question": "Is the hosepipe running constantly?", + "choice_a": "hand", + "choice_b": "bubbling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15001 + }, + { + "path": "Idling Truck at Loading Dock.wav", + "question": "is a dog barking?", + "choice_a": "none", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15002 + }, + { + "path": "Footsteps On Squeaky Wood Floor.wav", + "question": "Does the walking person wear shoes?", + "choice_a": "hand", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15003 + }, + { + "path": "meadow brook bees.wav", + "question": "What sweet product do the insects that make these sounds produce?", + "choice_a": "slowly", + "choice_b": "three", + "choice_c": "liquid", + "choice_d": "no", + "answer_gt": "liquid", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15004 + }, + { + "path": "Street Noise - Cars - Ball Bouncing indistinct voices.wav", + "question": "Does a vehicle go by?", + "choice_a": "spinner", + "choice_b": "forty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15005 + }, + { + "path": "Car Engine Idling.wav", + "question": "Do breaks squeal?", + "choice_a": "buzzsaw", + "choice_b": "clanking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15006 + }, + { + "path": "131227_strumyk_1.wav", + "question": "is the water falling lightly?", + "choice_a": "rain", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15007 + }, + { + "path": "Light to heavy Rain.wav", + "question": "Where is this?", + "choice_a": "sheep", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "ocean", + "answer_gt": "ocean", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15008 + }, + { + "path": "Wipers .wav", + "question": "Is a car accelerating?", + "choice_a": "bridge", + "choice_b": "zoo", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15009 + }, + { + "path": "walking indoors footsteps tap tap tapping foley.wav", + "question": "Are those footsteps?", + "choice_a": "drill", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15010 + }, + { + "path": "carpet_on_carpet.wav", + "question": "The hammering sound is interspersed with what other noise?", + "choice_a": "fifteen", + "choice_b": "sweeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "sweeping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15011 + }, + { + "path": "Household - Atmos - Wind Through Window.wav", + "question": "Is the wind blowing?", + "choice_a": "clapping", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15012 + }, + { + "path": "filling-ice-cup.wav", + "question": "Is the person pouring into a container?", + "choice_a": "chirp", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15013 + }, + { + "path": "chrrrckkkcc.wav", + "question": "Can a TV be heard in the background?", + "choice_a": "buzzsaw", + "choice_b": "barking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15014 + }, + { + "path": "Wind_Whistling_Dorm_Window.wav", + "question": "Does the pitch of the humming change?", + "choice_a": "owl", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15015 + }, + { + "path": "Typing 5 lines.wav", + "question": "Is there a ding?", + "choice_a": "twentyfive", + "choice_b": "sheep", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15016 + }, + { + "path": "Sink_Running2.wav", + "question": "Has something been turned on?", + "choice_a": "band", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15017 + }, + { + "path": "Air raid siren_rising.wav", + "question": "Are people screaming?", + "choice_a": "flood", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15018 + }, + { + "path": "Construction Zone.wav", + "question": "From which part of the car do the noise come from?", + "choice_a": "owl", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "engine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15019 + }, + { + "path": "Boom_Folie_NoiseOnGlass.wav", + "question": "How many people are speaking?", + "choice_a": "stormy", + "choice_b": "zoo", + "choice_c": "zero", + "choice_d": "no", + "answer_gt": "zero", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15020 + }, + { + "path": "indoors house ambient room tone distant neighbours 1.wav", + "question": "Are some sounds made by living beings?", + "choice_a": "grass", + "choice_b": "thunder", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15021 + }, + { + "path": "wind-sound-from-inside-car.wav", + "question": "Are there cars honking?", + "choice_a": "interested", + "choice_b": "bubbling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15022 + }, + { + "path": "Rain - 1.wav", + "question": "Is it from a tap?", + "choice_a": "honking", + "choice_b": "yes", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15023 + }, + { + "path": "second_floor_lav.wav", + "question": "What pitch tone is the object making?", + "choice_a": "snow", + "choice_b": "tearing", + "choice_c": "low", + "choice_d": "no", + "answer_gt": "low", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15024 + }, + { + "path": "water fountian-hollow sm med pitch.wav", + "question": "Is the person singing in the shower?", + "choice_a": "beeping", + "choice_b": "intercom", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15025 + }, + { + "path": "Footsteps on Rocky Terrain.wav", + "question": "What does it sound like?", + "choice_a": "working", + "choice_b": "cleaning", + "choice_c": "yes", + "choice_d": "walking", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15026 + }, + { + "path": "Water Faucet HQ Stereo.wav", + "question": "What item in a bathroom makes this sound?", + "choice_a": "rooster", + "choice_b": "drink", + "choice_c": "sink", + "choice_d": "no", + "answer_gt": "sink", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15027 + }, + { + "path": "Cityscape construction site 2 100304.wav", + "question": "How many cutting strokes does this person make ?", + "choice_a": "two", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15028 + }, + { + "path": "moving flipcharts after the meeting.wav", + "question": "Is a heavy object being moved?", + "choice_a": "nothing", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15029 + }, + { + "path": "Rain falling on a metal roof - 96 kHz _ 24 Bit.wav", + "question": "Where is this occuring?", + "choice_a": "outside", + "choice_b": "distortion", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "outside", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15030 + }, + { + "path": "Cat Meowing.wav", + "question": "What is growling?", + "choice_a": "yes", + "choice_b": "cat", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15031 + }, + { + "path": "adw018raw.wav", + "question": "What is ringing?", + "choice_a": "tearing", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "bell", + "answer_gt": "bell", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15032 + }, + { + "path": "Pag_Starigrad_crickets_birds_2.wav", + "question": "Is this outside?", + "choice_a": "fireworks", + "choice_b": "sweeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15033 + }, + { + "path": "night in the countryside.wav", + "question": "How many dogs are barking?", + "choice_a": "normal", + "choice_b": "bus", + "choice_c": "one", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15034 + }, + { + "path": "Siren Milan.wav", + "question": "How many sirens are going off?", + "choice_a": "nothing", + "choice_b": "crickets", + "choice_c": "yes", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15035 + }, + { + "path": "forest_ambience_chepachet_spring_1.wav", + "question": "Is the bird singing tranquil?", + "choice_a": "men", + "choice_b": "cars", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15036 + }, + { + "path": "Krankenwagen _ German Ambulances Passing by...wav", + "question": "What makes a siren sound?", + "choice_a": "drum", + "choice_b": "chicken", + "choice_c": "ambulance", + "choice_d": "no", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15037 + }, + { + "path": "Fantasy Ambience.wav", + "question": "Are there drums?", + "choice_a": "car", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15038 + }, + { + "path": "THE_RATT23_1.wav", + "question": "Is the person on a sidewalk?", + "choice_a": "leaves", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15039 + }, + { + "path": "STE-031 italian road.wav", + "question": "What is passing by at a regular and fast pace?", + "choice_a": "door", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "cars", + "answer_gt": "cars", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15040 + }, + { + "path": "Thunder_01.wav", + "question": "how many times does the thunder crack?", + "choice_a": "twice", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "twice", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15041 + }, + { + "path": "Stream # 2.wav", + "question": "Is tea being poured from a teapot?", + "choice_a": "hand", + "choice_b": "cabinet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15042 + }, + { + "path": "Footsteps Dress Shoes Wood Floor.wav", + "question": "How many steps does the person take?", + "choice_a": "forty", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "forty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15043 + }, + { + "path": "glissando1b.wav", + "question": "Does the bell sound change in volume?", + "choice_a": "twelve", + "choice_b": "clanking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15044 + }, + { + "path": "waves_1.wav", + "question": "what is moving?", + "choice_a": "cricket", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "waves", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15045 + }, + { + "path": "DoorSqueak.wav", + "question": "How many times does the door creak?", + "choice_a": "cars", + "choice_b": "five", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15046 + }, + { + "path": "BobKessler-Spinning Tin Top.wav", + "question": "What is being rolled down?", + "choice_a": "waves", + "choice_b": "coin", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "coin", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15047 + }, + { + "path": "footsteps on beach.wav", + "question": "Can people be heard in the cars driving by?", + "choice_a": "five", + "choice_b": "mixer", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15048 + }, + { + "path": "Beer Pong Sounds ball table and cups.wav", + "question": "Does a person cheer at the end?", + "choice_a": "radio", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15049 + }, + { + "path": "Old metal window open and close.wav", + "question": "Is there something moving along a surface?", + "choice_a": "rainfall", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15050 + }, + { + "path": "kitchen.wav", + "question": "Is something falling down the stairs?", + "choice_a": "bus", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15051 + }, + { + "path": "Cualquiera.wav", + "question": "Can the people be heard after the door is opened?", + "choice_a": "gas", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15052 + }, + { + "path": "Edit Radio .wav", + "question": "Is a person running?", + "choice_a": "clapping", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15053 + }, + { + "path": "wind and birds in the delta of the River Po 2.wav", + "question": "What is blowing?", + "choice_a": "one", + "choice_b": "wind", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "wind", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15054 + }, + { + "path": "Streatham Railway Station and on a Train.wav", + "question": "Is a cat scratching something?", + "choice_a": "car", + "choice_b": "clapping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15055 + }, + { + "path": "Horse_Hooves_Hard_Floor_Interior.wav", + "question": "does a person give out a yell?", + "choice_a": "fifteen", + "choice_b": "cards", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15056 + }, + { + "path": "machine1.wav", + "question": "Can people be heard fighting?", + "choice_a": "locust", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15057 + }, + { + "path": "Centurion Suburb Evening.wav", + "question": "Is there a bird chirping?", + "choice_a": "seagulls", + "choice_b": "leaves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15058 + }, + { + "path": "fireTruckFar NL 140109_00.wav", + "question": "Is there a crowd of people?", + "choice_a": "roof", + "choice_b": "never", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15059 + }, + { + "path": "bar crowd.wav", + "question": "how many questions are heard?", + "choice_a": "three", + "choice_b": "eight", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15060 + }, + { + "path": "Tools .wav", + "question": "Are people singing?", + "choice_a": "people", + "choice_b": "ocean", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15061 + }, + { + "path": "Face slap CsG.wav", + "question": "Does the slapping sound happen more than once?", + "choice_a": "machine", + "choice_b": "twentyfive", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15062 + }, + { + "path": "Hitting baseball w. wooden bat.wav", + "question": "Can an object hitting a ball be heard?", + "choice_a": "spinner", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15063 + }, + { + "path": "Rainforest Morning Chorus.wav", + "question": "What are communicating with each other?", + "choice_a": "grass", + "choice_b": "speaking", + "choice_c": "birds", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15064 + }, + { + "path": "Metal clatter drop.wav", + "question": "Is a rooster crowing?", + "choice_a": "bike", + "choice_b": "zoo", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15065 + }, + { + "path": "Plaza_de_la_Revolucion_risa.wav", + "question": "Are there people laughing?", + "choice_a": "bear", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15066 + }, + { + "path": "train.wav", + "question": "does the speed of the vehicle stay the same as time goes on?", + "choice_a": "low", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15067 + }, + { + "path": "CR FunnyMachine.wav", + "question": "Does the machine come to a stop?", + "choice_a": "rod", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15068 + }, + { + "path": "air bubbles on the surface of the water.wav", + "question": "What type of water device is in operation?", + "choice_a": "drum", + "choice_b": "grass", + "choice_c": "yes", + "choice_d": "tap", + "answer_gt": "tap", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15069 + }, + { + "path": "FOLEY_Ext_Garbage_Hauling_001.wav", + "question": "What was heard at the start of the audio recording?", + "choice_a": "walking", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "motor", + "answer_gt": "motor", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15070 + }, + { + "path": "car acceleration sequence.wav", + "question": "is there a person talking in the car?", + "choice_a": "cars", + "choice_b": "train", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15071 + }, + { + "path": "Blackbird sounds.wav", + "question": "Is a violin being played?", + "choice_a": "roof", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15072 + }, + { + "path": "Howling Storm.wav", + "question": "Is the sound intensity increasing before stabilising?", + "choice_a": "three", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15073 + }, + { + "path": "Flipping Coin Can.wav", + "question": "Are objects colliding with each other?", + "choice_a": "road", + "choice_b": "raining", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15074 + }, + { + "path": "Short Hailstorm.wav", + "question": "What is pouring down in the sound?", + "choice_a": "gas", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15075 + }, + { + "path": "Wipers .wav", + "question": "How many times do the car wipers go across the windscreen?", + "choice_a": "heavy", + "choice_b": "normal", + "choice_c": "twelve", + "choice_d": "no", + "answer_gt": "twelve", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15076 + }, + { + "path": "Centurion Suburb Evening.wav", + "question": "Is there a bird chirping?", + "choice_a": "party", + "choice_b": "clapping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15077 + }, + { + "path": "Walking alongside the road.wav", + "question": "Is the person eating something?", + "choice_a": "heavy", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15078 + }, + { + "path": "windup_dino_slow.wav", + "question": "What is the person doing?", + "choice_a": "people", + "choice_b": "raining", + "choice_c": "driving", + "choice_d": "no", + "answer_gt": "driving", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15079 + }, + { + "path": "Wind-up Toy Motorbike SFX.wav", + "question": "If this sound was coming from a toy, what would that toy be called?", + "choice_a": "beeping", + "choice_b": "whispering", + "choice_c": "spinner", + "choice_d": "no", + "answer_gt": "spinner", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15080 + }, + { + "path": "RemoteControl.Antique.Zenith.wav", + "question": "Is it a voice?", + "choice_a": "cleaning", + "choice_b": "nothing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15081 + }, + { + "path": "faucet3.wav", + "question": "Is anything being broken?", + "choice_a": "stops", + "choice_b": "six", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15082 + }, + { + "path": "Fairground 2 Ghost ride.wav", + "question": "Is this inside a haunted house?", + "choice_a": "fifteen", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15083 + }, + { + "path": "Senseo_boil_norm.wav", + "question": "does someone drop anything?", + "choice_a": "water", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15084 + }, + { + "path": "julies media.wav", + "question": "How many times did the siren go off?", + "choice_a": "static", + "choice_b": "music", + "choice_c": "ten", + "choice_d": "no", + "answer_gt": "ten", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15085 + }, + { + "path": "babbling brook 2.wav", + "question": "Is someone sipping from a cup?", + "choice_a": "talking", + "choice_b": "party", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15086 + }, + { + "path": "nxSample010.wav", + "question": "Is water running?", + "choice_a": "bubbling", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15087 + }, + { + "path": "basement-stairs.wav", + "question": "Is it a staccato sound?", + "choice_a": "screeching", + "choice_b": "medium", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15088 + }, + { + "path": "quedlinburg castle.wav", + "question": "Is there only one person?", + "choice_a": "locust", + "choice_b": "cars", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15089 + }, + { + "path": "MISC_Int_Cat_Purring_002.wav", + "question": "What is being used to groom the cat?", + "choice_a": "band", + "choice_b": "static", + "choice_c": "yes", + "choice_d": "brush", + "answer_gt": "brush", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15090 + }, + { + "path": "water slushing moderate speed.wav", + "question": "What liquid is being played with?", + "choice_a": "radio", + "choice_b": "rainfall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rainfall", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15091 + }, + { + "path": "treefrogs.wav", + "question": "Is the steam engine hissing?", + "choice_a": "zero", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15092 + }, + { + "path": "Night Frogs.wav", + "question": "Is it night time?", + "choice_a": "sweeping", + "choice_b": "sink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15093 + }, + { + "path": "amazon 04.wav", + "question": "Is it a crying baby?", + "choice_a": "medium", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15094 + }, + { + "path": "Chime-Ball.wav", + "question": "Is the sound staticky?", + "choice_a": "drill", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15095 + }, + { + "path": "Tools .wav", + "question": "Are metallic objects being moved?", + "choice_a": "once", + "choice_b": "twenty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15096 + }, + { + "path": "Rain and Storm.wav", + "question": "Is the rain hitting an umbrella?", + "choice_a": "hands", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15097 + }, + { + "path": "pr#6F9A9E.wav", + "question": "does the item break?", + "choice_a": "paper", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15098 + }, + { + "path": "Spring Birds Raw _New Jersey_.wav", + "question": "What domestic animal likes to catch these creatures?", + "choice_a": "seven", + "choice_b": "wet", + "choice_c": "birds", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15099 + }, + { + "path": "Flipping Coin Can.wav", + "question": "Is this water flowing down a river?", + "choice_a": "screeching", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15100 + }, + { + "path": "drunk_teenagers_1.wav", + "question": "Is this indoors?", + "choice_a": "two", + "choice_b": "five", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15101 + }, + { + "path": "Teluk Nipah 01.wav", + "question": "Is there something in the water?", + "choice_a": "cards", + "choice_b": "seesaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15102 + }, + { + "path": "20090407.cricket.real.close.wav", + "question": "When is the noise the quietest?", + "choice_a": "twenty", + "choice_b": "nothing", + "choice_c": "none", + "choice_d": "no", + "answer_gt": "none", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15103 + }, + { + "path": "20100801.wharf.silence.night.wav", + "question": "What is the animal doing?", + "choice_a": "ten", + "choice_b": "birds", + "choice_c": "barking", + "choice_d": "no", + "answer_gt": "barking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15104 + }, + { + "path": "Ext-amb_park_late-fall_distant-gun-shot_Distant-Child-shouting.wav", + "question": "What subtle sound can be heard once toward the end?", + "choice_a": "medium", + "choice_b": "eating", + "choice_c": "yes", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15105 + }, + { + "path": "Night in nature.wav", + "question": "what animal could make this noise?", + "choice_a": "none", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15106 + }, + { + "path": "20090407.cricket.real.close.wav", + "question": "Is a sound a human would make?", + "choice_a": "swimming", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15107 + }, + { + "path": "Footsteps Dress Shoes Wood Floor.wav", + "question": "What is the activity that's taking place?", + "choice_a": "siren", + "choice_b": "engine", + "choice_c": "walking", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15108 + }, + { + "path": "Slushing in mouth.wav", + "question": "Does the man gulp down the drink?", + "choice_a": "twentythree", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15109 + }, + { + "path": "indoors dorm dormitory ambient room tone distant traffic in street.wav", + "question": "How many machines are being used?", + "choice_a": "announcement", + "choice_b": "rainfall", + "choice_c": "yes", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15110 + }, + { + "path": "Little sreet behind a terrasse cafe.wav", + "question": "Are there animals nearby?", + "choice_a": "canvas", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15111 + }, + { + "path": "corneille_city01.wav", + "question": "Is there traffic in the background?", + "choice_a": "flood", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15112 + }, + { + "path": "wheaten field.wav", + "question": "Can people be heard yelling?", + "choice_a": "heavy", + "choice_b": "buzzing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15113 + }, + { + "path": "April dawn chorus Sydenham Hill.wav", + "question": "What type of animal is heard?", + "choice_a": "twentyfive", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15114 + }, + { + "path": "crowdfree.wav", + "question": "Are there birds chirping?", + "choice_a": "showering", + "choice_b": "cricket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15115 + }, + { + "path": "Thunder Rain Cars Driving By.wav", + "question": "How is the weather?", + "choice_a": "bubbling", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "stormy", + "answer_gt": "stormy", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15116 + }, + { + "path": "Kitchen faucet running fast and slow and filling glass of water.wav", + "question": "what is being filled with water?", + "choice_a": "sink", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "sink", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15117 + }, + { + "path": "Cars_bridge.wav", + "question": "are horns honking?", + "choice_a": "door", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15118 + }, + { + "path": "20091224.bells.01.wav", + "question": "Is the sound being made without the help of a person?", + "choice_a": "clanking", + "choice_b": "electronic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15119 + }, + { + "path": "market.wav", + "question": "How many different people are talking?", + "choice_a": "five", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15120 + }, + { + "path": "cupboard door squeaks.wav", + "question": "What is being opened and closed?", + "choice_a": "cards", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "door", + "answer_gt": "door", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15121 + }, + { + "path": "the swarm v31m3.wav", + "question": "What happens to the sound level at the end?", + "choice_a": "cleaning", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "stops", + "answer_gt": "stops", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15122 + }, + { + "path": "Frogs in town.wav", + "question": "do any humans talk?", + "choice_a": "mixer", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15123 + }, + { + "path": "Japan_Tokyo_Shinjuku_Street_Promoter_Yelling_City.wav", + "question": "What can be heard going by in the background?", + "choice_a": "car", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "car", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15124 + }, + { + "path": "UnionStation06OutBack_BusyOutside.wav", + "question": "Is there a person in the area?", + "choice_a": "locust", + "choice_b": "swimming", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15125 + }, + { + "path": "Garden Birds 3.wav", + "question": "What are these creatures?", + "choice_a": "sheep", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15126 + }, + { + "path": "Rain - 1.wav", + "question": "does the water sound get louder?", + "choice_a": "brush", + "choice_b": "ground", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15127 + }, + { + "path": "public.wav", + "question": "Are any birds making noise?", + "choice_a": "static", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15128 + }, + { + "path": "background of the side streets of Rhodes, scooter, tourists French and American, grinder.wav", + "question": "What is being cut in the recording?", + "choice_a": "screeching", + "choice_b": "fifteen", + "choice_c": "tree", + "choice_d": "no", + "answer_gt": "tree", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15129 + }, + { + "path": "aftertherain.wav", + "question": "What are the vehicles likely traveling on?", + "choice_a": "crows", + "choice_b": "bridge", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bridge", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15130 + }, + { + "path": "sparrows.wav", + "question": "Is there more than one bird?", + "choice_a": "start", + "choice_b": "rainfall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15131 + }, + { + "path": "Coffeehouse Ambience Burlington VT 0112xx.wav", + "question": "Is the person alone?", + "choice_a": "washing", + "choice_b": "glass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15132 + }, + { + "path": "Train Pass Koln.wav", + "question": "what type of vehicle can be heard?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15133 + }, + { + "path": "Silencyo_CC_Field_Cicadas_French Children Playing_Wind_Dog_Distant Traffic.wav", + "question": "Was this person inside?", + "choice_a": "liquid", + "choice_b": "canvas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15134 + }, + { + "path": "clanking lid.wav", + "question": "what is the metal object doing?", + "choice_a": "distortion", + "choice_b": "clanging", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "clanging", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15135 + }, + { + "path": "auto-rickshaw-trip.wav", + "question": "Is there more than one vehicle?", + "choice_a": "lock", + "choice_b": "siren", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15136 + }, + { + "path": "metal_bowls_altered.wav", + "question": "What kind of pitch does the sound have?", + "choice_a": "medium", + "choice_b": "sand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "medium", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15137 + }, + { + "path": "squeaky metal swing.wav", + "question": "Did someone slam on the brakes?", + "choice_a": "wet", + "choice_b": "static", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15138 + }, + { + "path": "LOUD THUNDER - WITH RAIN HITTING UMBRELLA.wav", + "question": "How many birds are chirping?", + "choice_a": "gas", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15139 + }, + { + "path": "Mulholland Memorial Fountain Los Angeles.wav", + "question": "Does the static noise change to something else?", + "choice_a": "women", + "choice_b": "bucket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15140 + }, + { + "path": "AM static.wav", + "question": "Is that a human voice?", + "choice_a": "drum", + "choice_b": "walking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15141 + }, + { + "path": "Kaffemaschine_1.wav", + "question": "What is the sound the machine makes?", + "choice_a": "drilling", + "choice_b": "twelve", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "drilling", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15142 + }, + { + "path": "Garden Birds 3.wav", + "question": "Is this outside?", + "choice_a": "window", + "choice_b": "bear", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15143 + }, + { + "path": "Koeien, R4 en riet Lichterveldestraat.wav", + "question": "How many tires do these vehicles, when passenger sized, typically have?", + "choice_a": "four", + "choice_b": "showering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "four", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15144 + }, + { + "path": "WindInPylons.wav", + "question": "Is this inside a train ?", + "choice_a": "train", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15145 + }, + { + "path": "Cardiff Bay fireworks.wav", + "question": "What is being celebrated?", + "choice_a": "bus", + "choice_b": "wrench", + "choice_c": "party", + "choice_d": "no", + "answer_gt": "party", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15146 + }, + { + "path": "Old metal window open and close.wav", + "question": "How many times can the scraping sound be heard?", + "choice_a": "locust", + "choice_b": "normal", + "choice_c": "yes", + "choice_d": "six", + "answer_gt": "six", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15147 + }, + { + "path": "WavesOnTheShore.wav", + "question": "What can be heard being moved?", + "choice_a": "chirping", + "choice_b": "chain", + "choice_c": "yes", + "choice_d": "water", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15148 + }, + { + "path": "FootstepsOnWoodFloor.wav", + "question": "Is a shoe being used?", + "choice_a": "grass", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15149 + }, + { + "path": "STE-031 italian road.wav", + "question": "Is there lots of spray?", + "choice_a": "chirp", + "choice_b": "leaves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15150 + }, + { + "path": "the swarm v31m3.wav", + "question": "Is the fan on high?", + "choice_a": "four", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15151 + }, + { + "path": "wind and birds in the delta of the River Po 2.wav", + "question": "What are the birds doing?", + "choice_a": "working", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "chirping", + "answer_gt": "chirping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15152 + }, + { + "path": "knocking on a window or glass.wav", + "question": "How many knocks can be heard?", + "choice_a": "twenty", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "twenty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15153 + }, + { + "path": "20100801.wharf.silence.night.wav", + "question": "Is the animal a cat?", + "choice_a": "people", + "choice_b": "interested", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15154 + }, + { + "path": "outdoors ambient village bird distant neighbours children car.wav", + "question": "Are people speaking?", + "choice_a": "short", + "choice_b": "buzzing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15155 + }, + { + "path": "Fast stream _ small river.wav", + "question": "Is the liquid moving quickly?", + "choice_a": "yes", + "choice_b": "loudspeaker", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15156 + }, + { + "path": "Hyeres street voices ambience f.wav", + "question": "Does someone cough?", + "choice_a": "cabinet", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15157 + }, + { + "path": "Marker Writing on Paper.wav", + "question": "What is the person writing on?", + "choice_a": "stormy", + "choice_b": "cow", + "choice_c": "thirty", + "choice_d": "no", + "answer_gt": "thirty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15158 + }, + { + "path": "cupboard door squeaks.wav", + "question": "Is wood making this noise?", + "choice_a": "once", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15159 + }, + { + "path": "knocking on a window or glass.wav", + "question": "Is this an animal?", + "choice_a": "raining", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15160 + }, + { + "path": "JM_HOME&OFFICE_Shower 01 - Taking a shower.wav", + "question": "Are there people talking?", + "choice_a": "cabinet", + "choice_b": "heavy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15161 + }, + { + "path": "car acceleration sequence.wav", + "question": "Is a car being driven?", + "choice_a": "twentyfive", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15162 + }, + { + "path": "quedlinburg castle.wav", + "question": "Are people walking?", + "choice_a": "coin", + "choice_b": "twentythree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15163 + }, + { + "path": "FOLEY_Ext_Garbage_Hauling_001.wav", + "question": "Did it sound like there was something rolling around on a metal floor?", + "choice_a": "saw", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15164 + }, + { + "path": "crumpleTissuePaper.wav", + "question": "Is something being crinkled?", + "choice_a": "roof", + "choice_b": "heavy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15165 + }, + { + "path": "windroar_constant_1m12s.wav", + "question": "How heavy is the rainfall?", + "choice_a": "honking", + "choice_b": "big", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "big", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15166 + }, + { + "path": "WavesOnTheShore.wav", + "question": "Can more that a dozen people be heard?", + "choice_a": "zero", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15167 + }, + { + "path": "Teluk Nipah 01.wav", + "question": "Is there something in the water?", + "choice_a": "announcement", + "choice_b": "radio", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15168 + }, + { + "path": "audience final applause 01.wav", + "question": "what are people using to make sounds?", + "choice_a": "hands", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "hands", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15169 + }, + { + "path": "ROLDAN_ANDRES_GEO_FONIA_AGUA_DUCHA.wav", + "question": "What is the water doing?", + "choice_a": "motor", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "falls", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15170 + }, + { + "path": "Conversacion Punjabi.wav", + "question": "how many times is the tap heard?", + "choice_a": "running", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "zero", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15171 + }, + { + "path": "Gentle rain outside balcony street noise.wav", + "question": "Is there more sounds than just an engine?", + "choice_a": "rainfall", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15172 + }, + { + "path": "metal rain.wav", + "question": "Is a child singing?", + "choice_a": "thirty", + "choice_b": "chirping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15173 + }, + { + "path": "Water in a canal.wav", + "question": "Does the sound come from an enclosed environment?", + "choice_a": "fifteen", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15174 + }, + { + "path": "Organic sound.wav", + "question": "Is someone moving something around?", + "choice_a": "roof", + "choice_b": "hand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15175 + }, + { + "path": "Ambience - St Kilda Beach - waves lapping rocks, people nearby, seagulls.wav", + "question": "Is there wading?", + "choice_a": "buzzsaw", + "choice_b": "three", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15176 + }, + { + "path": "Steps Indoor medium soft Shoe Sole accompanying wooden Floor hollow Big Room 5mx10mx6m.wav", + "question": "What's the person doing?", + "choice_a": "walking", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15177 + }, + { + "path": "Single cricket chirping during a summer evening in the city _with traffic noise_.wav", + "question": "Is someone pouring out a liquid?", + "choice_a": "rainy", + "choice_b": "loudspeaker", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15178 + }, + { + "path": "Bubbles water.wav", + "question": "Is something burning?", + "choice_a": "bird", + "choice_b": "yes", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15179 + }, + { + "path": "LOUD THUNDER - WITH RAIN HITTING UMBRELLA.wav", + "question": "Do the noises come from a machine?", + "choice_a": "chain", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15180 + }, + { + "path": "Drumming on a wine glass.wav", + "question": "Are there other sounds besides the tapping one?", + "choice_a": "drilling", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15181 + }, + { + "path": "Thunder_01.wav", + "question": "How is the weather now?", + "choice_a": "bus", + "choice_b": "crickets", + "choice_c": "raining", + "choice_d": "no", + "answer_gt": "raining", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15182 + }, + { + "path": "Wet_Soggy_Squishy_Footsteps.wav", + "question": "How many times is the object squished?", + "choice_a": "eight", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "seven", + "answer_gt": "seven", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15183 + }, + { + "path": "divide lake.wav", + "question": "Is it outside?", + "choice_a": "swimming", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15184 + }, + { + "path": "chrrrckkkcc.wav", + "question": "How many times in the clanging noise made?", + "choice_a": "three", + "choice_b": "crows", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15185 + }, + { + "path": "20061124ParadeCHS.wav", + "question": "Is this outside?", + "choice_a": "spinner", + "choice_b": "cabinet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15186 + }, + { + "path": "20130327_valparaiso.traffic.02.wav", + "question": "Is someone screaming?", + "choice_a": "closet", + "choice_b": "small", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15187 + }, + { + "path": "restaurant wood floor.wav", + "question": "Can eating utensils be heard clinking against plates?", + "choice_a": "twentyfive", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15188 + }, + { + "path": "Mariehamn_frogs.wav", + "question": "is the sound indoors?", + "choice_a": "chopping", + "choice_b": "five", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15189 + }, + { + "path": "Caltrain Pushing Caltrain.wav", + "question": "Does the horn blow at least once?", + "choice_a": "owl", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15190 + }, + { + "path": "Train Pass Koln.wav", + "question": "Is this at the airport?", + "choice_a": "cabinet", + "choice_b": "low", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15191 + }, + { + "path": "steps_snow.wav", + "question": "Is there more than one person marching?", + "choice_a": "rooster", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15192 + }, + { + "path": "Storm sirens with dog bark at end 050627 24 bit.wav", + "question": "Is it raining?", + "choice_a": "clapping", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15193 + }, + { + "path": "abandoned-ballroom-radiators.wav", + "question": "Is this a marching band?", + "choice_a": "metal", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15194 + }, + { + "path": "Mulholland Memorial Fountain Los Angeles.wav", + "question": "Does this sound like static?", + "choice_a": "tank", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15195 + }, + { + "path": "Page turns and book close_open.wav", + "question": "Is it a repetitive sound?", + "choice_a": "men", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15196 + }, + { + "path": "Serving Water Quickly.wav", + "question": "Is someone setting silverware against the table?", + "choice_a": "fireworks", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15197 + }, + { + "path": "Walking in Grass in Evening with Loud Bird.wav", + "question": "is the person walking on grass?", + "choice_a": "crows", + "choice_b": "open", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15198 + }, + { + "path": "InsectsSummerSwellingAug16th2013.wav", + "question": "Is it a conveyor belt?", + "choice_a": "drilling", + "choice_b": "fifteen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15199 + }, + { + "path": "Quill pen writing on hard paper various speed.wav", + "question": "Is something being carved?", + "choice_a": "zoo", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15200 + }, + { + "path": "HarleyDavidson.wav", + "question": "Is a motorcycle rumbling?", + "choice_a": "cutter", + "choice_b": "drum", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15201 + }, + { + "path": "Arch Leaf.wav", + "question": "What are they walking on?", + "choice_a": "grass", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "grass", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15202 + }, + { + "path": "Autumnal Ambient 24 Bits 48 Khz.wav", + "question": "Is this in nature?", + "choice_a": "snow", + "choice_b": "eating", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15203 + }, + { + "path": "Heavy Rain 1.wav", + "question": "Is the sound of rain constant?", + "choice_a": "metal", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15204 + }, + { + "path": "Walking On Dry Leaves Normalised.wav", + "question": "What is happening here?", + "choice_a": "thirty", + "choice_b": "thunder", + "choice_c": "yes", + "choice_d": "cleaning", + "answer_gt": "cleaning", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15205 + }, + { + "path": "3trump.wav", + "question": "Is there more than one instrument being played?", + "choice_a": "barking", + "choice_b": "metal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15206 + }, + { + "path": "T24_home_dishwasher_cycle change.wav", + "question": "What is making this sound?", + "choice_a": "machine", + "choice_b": "zoo", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "machine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15207 + }, + { + "path": "glass d.wav", + "question": "Did someone yawn?", + "choice_a": "cricket", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15208 + }, + { + "path": "mediterranean_sea_porticcio.wav", + "question": "is the water moving?", + "choice_a": "spinner", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15209 + }, + { + "path": "AM static.wav", + "question": "Is that a human voice?", + "choice_a": "lock", + "choice_b": "normal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15210 + }, + { + "path": "sparrows.wav", + "question": "Are the birds making a soft sound?", + "choice_a": "sink", + "choice_b": "zoo", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15211 + }, + { + "path": "second_floor_lav.wav", + "question": "is the sound constant?", + "choice_a": "keys", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15212 + }, + { + "path": "RKeaton_EMF366_12_Tearing Thick Paper.wav", + "question": "What is the person doing to the paper?", + "choice_a": "party", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "tearing", + "answer_gt": "tearing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15213 + }, + { + "path": "20091217.18.chains.wav", + "question": "Does a police car drive past?", + "choice_a": "hand", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15214 + }, + { + "path": "worktoilet.wav", + "question": "Is there water?", + "choice_a": "flood", + "choice_b": "fireworks", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15215 + }, + { + "path": "Small Falling Water Onto Stones.wav", + "question": "Is there a lot of water?", + "choice_a": "twentythree", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15216 + }, + { + "path": "Thunder Outside.wav", + "question": "Are people running away from the thunder?", + "choice_a": "thunder", + "choice_b": "locust", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15217 + }, + { + "path": "Large Splashes.wav", + "question": "How many splashes are there?", + "choice_a": "stops", + "choice_b": "leaves", + "choice_c": "five", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15218 + }, + { + "path": "winter wren wind leaves.wav", + "question": "What sound does the bird make?", + "choice_a": "drill", + "choice_b": "water", + "choice_c": "chirp", + "choice_d": "no", + "answer_gt": "chirp", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15219 + }, + { + "path": "Bicycle bell.wav", + "question": "Is anyone talking?", + "choice_a": "chirping", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15220 + }, + { + "path": "stone_well.wav", + "question": "is there a consistent sound of cricket like things?", + "choice_a": "stops", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15221 + }, + { + "path": "AMB_COLE.wav", + "question": "Is this a small gathering?", + "choice_a": "six", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15222 + }, + { + "path": "SonicSnap_GPSUK_Cockerel.wav", + "question": "Is this recording a repetition of just one distinctive noise?", + "choice_a": "rainy", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15223 + }, + { + "path": "carpet_on_carpet.wav", + "question": "Is someone building something?", + "choice_a": "party", + "choice_b": "cat", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15224 + }, + { + "path": "cricket chirp.wav", + "question": "Are frogs croaking?", + "choice_a": "hen", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15225 + }, + { + "path": "metal rain.wav", + "question": "Is a child singing?", + "choice_a": "motor", + "choice_b": "three", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15226 + }, + { + "path": "crowdfree.wav", + "question": "Is there honking?", + "choice_a": "cricket", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15227 + }, + { + "path": "110709_05 goma exhibit.wav", + "question": "Are there people performing?", + "choice_a": "five", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15228 + }, + { + "path": "20060523.grassland.wav", + "question": "Is there wildlife close?", + "choice_a": "closet", + "choice_b": "people", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15229 + }, + { + "path": "LakeSide2.wav", + "question": "How many people make any noise?", + "choice_a": "drilling", + "choice_b": "birds", + "choice_c": "two", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15230 + }, + { + "path": "Traffic Light.wav", + "question": "What are the people doing in the background?", + "choice_a": "talking", + "choice_b": "flood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "talking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15231 + }, + { + "path": "neumatico.wav", + "question": "Is something being hammered?", + "choice_a": "heavy", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15232 + }, + { + "path": "LA Rain.wav", + "question": "Is there water?", + "choice_a": "brush", + "choice_b": "raining", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15233 + }, + { + "path": "peanutFarmDawnShort.wav", + "question": "What sound is the bird making?", + "choice_a": "chirp", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "chirp", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15234 + }, + { + "path": "Water_Drops_Falling.wav", + "question": "What is making the crackling noise?", + "choice_a": "wood", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "wood", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15235 + }, + { + "path": "Air raid siren_rising.wav", + "question": "Are the birds singing?", + "choice_a": "coin", + "choice_b": "chicken", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15236 + }, + { + "path": "Cooking rice.wav", + "question": "Is a bus honking?", + "choice_a": "hen", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15237 + }, + { + "path": "forest.wav", + "question": "What is the loud sound in distance ?", + "choice_a": "machine", + "choice_b": "canvas", + "choice_c": "wind", + "choice_d": "no", + "answer_gt": "wind", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15238 + }, + { + "path": "20081130_walking_in_snow.wav", + "question": "What are they walking in?", + "choice_a": "radio", + "choice_b": "twentythree", + "choice_c": "sand", + "choice_d": "no", + "answer_gt": "sand", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15239 + }, + { + "path": "diesel_train_comes.wav", + "question": "Are there multiple vehicles?", + "choice_a": "crickets", + "choice_b": "none", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15240 + }, + { + "path": "Gravel_Sand Walking 1.wav", + "question": "Is there more than one person?", + "choice_a": "cricket", + "choice_b": "siren", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15241 + }, + { + "path": "bathroom drain 1.wav", + "question": "Is there a spring of water?", + "choice_a": "people", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15242 + }, + { + "path": "broken comms2.wav", + "question": "What happens to the communication signal?", + "choice_a": "sweeping", + "choice_b": "machine", + "choice_c": "distortion", + "choice_d": "no", + "answer_gt": "distortion", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15243 + }, + { + "path": "Inner City Bees.wav", + "question": "Are humans heard?", + "choice_a": "speaking", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15244 + }, + { + "path": "bar crowd.wav", + "question": "how many questions are heard?", + "choice_a": "three", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15245 + }, + { + "path": "channel 2 now concludes its broadcast day.wav", + "question": "What kind of building might have machinery that makes this noise?", + "choice_a": "building", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "building", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15246 + }, + { + "path": "metal-bell-percussion.wav", + "question": "Are there occasional periods of silence?", + "choice_a": "women", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15247 + }, + { + "path": "cowshed.wav", + "question": "What animal is making the noise?", + "choice_a": "police", + "choice_b": "bus", + "choice_c": "cow", + "choice_d": "no", + "answer_gt": "cow", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15248 + }, + { + "path": "AMB_COLE.wav", + "question": "Are people speaking one at a time?", + "choice_a": "flood", + "choice_b": "brush", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15249 + }, + { + "path": "down stars running 3.wav", + "question": "What kind of noise is it?", + "choice_a": "steps", + "choice_b": "twice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "steps", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15250 + }, + { + "path": "Train sound.wav", + "question": "is something moving?", + "choice_a": "walking", + "choice_b": "people", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15251 + }, + { + "path": "Cooking on Gas.wav", + "question": "Does it get louder?", + "choice_a": "tearing", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15252 + }, + { + "path": "Jumping onto a hard floor with shoes and some walking sounds.wav", + "question": "What is scraping on the floor?", + "choice_a": "heavy", + "choice_b": "flushing", + "choice_c": "paper", + "choice_d": "no", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15253 + }, + { + "path": "Watering Can.wav", + "question": "Is the liquid moving from one location to another?", + "choice_a": "driller", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15254 + }, + { + "path": "Chicharra1.wav", + "question": "Can someone be heard shifting around?", + "choice_a": "drink", + "choice_b": "hall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15255 + }, + { + "path": "Llantas_rechinando.wav", + "question": "How many times is a screeching sound made?", + "choice_a": "chain", + "choice_b": "sand", + "choice_c": "six", + "choice_d": "no", + "answer_gt": "six", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15256 + }, + { + "path": "AMB_COLE.wav", + "question": "Are people speaking one at a time?", + "choice_a": "cleaning", + "choice_b": "hall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15257 + }, + { + "path": "Rain and Storm.wav", + "question": "Does it sound like it's raining hard there?", + "choice_a": "rooster", + "choice_b": "three", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15258 + }, + { + "path": "miniature goats and sheep.wav", + "question": "What animal is making the noise?", + "choice_a": "once", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "sheep", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15259 + }, + { + "path": "Cityscape 04 090617.wav", + "question": "Are people operating vehicles?", + "choice_a": "bike", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15260 + }, + { + "path": "bellaromani.wav", + "question": "How many times is the bell struck?", + "choice_a": "twentyfive", + "choice_b": "seven", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "seven", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15261 + }, + { + "path": "JM_HOME&OFFICE_Shower 01 - Taking a shower.wav", + "question": "Are there people talking?", + "choice_a": "cricket", + "choice_b": "no", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15262 + }, + { + "path": "Parking Garage - Ambiance, Electrical Hum 1.wav", + "question": "What is the sound that heard called?", + "choice_a": "buzzing", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "buzzing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15263 + }, + { + "path": "Wisper1.wav", + "question": "What is the person doing?", + "choice_a": "normal", + "choice_b": "whispering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "whispering", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15264 + }, + { + "path": "Conversacion Punjabi.wav", + "question": "What are the people doing?", + "choice_a": "zero", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "talking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15265 + }, + { + "path": "fan_2_300513.wav", + "question": "Are there people chatting?", + "choice_a": "men", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15266 + }, + { + "path": "pencil on paper.wav", + "question": "Is someone writing on a chalkboard?", + "choice_a": "screeching", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15267 + }, + { + "path": "indoors dorm dormitory ambient room tone distant traffic in street.wav", + "question": "Are there people running?", + "choice_a": "showering", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15268 + }, + { + "path": "draining board metal drip on metal.wav", + "question": "does the dripping water make the same sound throughout?", + "choice_a": "outside", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15269 + }, + { + "path": "faucet3.wav", + "question": "does the sound get heavier then die down?", + "choice_a": "wind", + "choice_b": "train", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15270 + }, + { + "path": "car-radio-am-noise2.wav", + "question": "Is the sound soothing to sleep to?", + "choice_a": "birds", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15271 + }, + { + "path": "Chicharra1.wav", + "question": "Can someone be heard shifting around?", + "choice_a": "flood", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15272 + }, + { + "path": "Subping03.wav", + "question": "What machine creates this noise?", + "choice_a": "rainfall", + "choice_b": "building", + "choice_c": "electronic", + "choice_d": "no", + "answer_gt": "electronic", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15273 + }, + { + "path": "background of the side streets of Rhodes, scooter, tourists French and American, grinder.wav", + "question": "Is anybody playing basketball?", + "choice_a": "glass", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15274 + }, + { + "path": "20090407.cricket.real.close.wav", + "question": "Does the noise stop in the middle?", + "choice_a": "clanging", + "choice_b": "sweeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15275 + }, + { + "path": "Naoshima-squeeking.wav", + "question": "Does the motor run steadily?", + "choice_a": "eight", + "choice_b": "bell", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15276 + }, + { + "path": "corneille_city01.wav", + "question": "How many birds are there?", + "choice_a": "water", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15277 + }, + { + "path": "Big_Roundabout_Traffic.wav", + "question": "Is this a vehicle?", + "choice_a": "hall", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15278 + }, + { + "path": "d0_drips_04.wav", + "question": "What is dripping into glass?", + "choice_a": "tap", + "choice_b": "windy", + "choice_c": "water", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15279 + }, + { + "path": "Hamamatsu-traffic-light-1.wav", + "question": "Does the ringing stop?", + "choice_a": "cutter", + "choice_b": "drum", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15280 + }, + { + "path": "Davis.wav", + "question": "Is there precipitation today?", + "choice_a": "engine", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15281 + }, + { + "path": "100121.wav", + "question": "How many separate explosion sounds were there?", + "choice_a": "three", + "choice_b": "cutter", + "choice_c": "one", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15282 + }, + { + "path": "Fryers Forest - Powerful Owl _Ninox Stenua_.wav", + "question": "Does the audio take place indoors?", + "choice_a": "sweeping", + "choice_b": "yes", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15283 + }, + { + "path": "Buddhist Bells.wav", + "question": "Are the reverberations are short in duration?", + "choice_a": "windy", + "choice_b": "sand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15284 + }, + { + "path": "sparrows.wav", + "question": "What is the bird doing?", + "choice_a": "chirping", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "chirping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15285 + }, + { + "path": "night in the countryside.wav", + "question": "How many dogs are barking?", + "choice_a": "announcement", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15286 + }, + { + "path": "stone_well.wav", + "question": "Is there more than one type of sound?", + "choice_a": "cabinet", + "choice_b": "drum", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15287 + }, + { + "path": "Autumnal Ambient 24 Bits 48 Khz.wav", + "question": "Is this in nature?", + "choice_a": "heavy", + "choice_b": "cards", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15288 + }, + { + "path": "Kaffemaschine_1.wav", + "question": "What is been used?", + "choice_a": "talking", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "saw", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15289 + }, + { + "path": "mechanical.wav", + "question": "are many machines being worked?", + "choice_a": "sound", + "choice_b": "speaking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15290 + }, + { + "path": "Big_Roundabout_Traffic.wav", + "question": "How many vehicles can be heard?", + "choice_a": "five", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15291 + }, + { + "path": "carnival_parade_cologne_1.wav", + "question": "What is heard behind the man speaking?", + "choice_a": "party", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "music", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15292 + }, + { + "path": "0221 Bar_terrace.wav", + "question": "Are there people around?", + "choice_a": "cow", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15293 + }, + { + "path": "LightRaininPinesMarch302013.wav", + "question": "What is the rain hitting?", + "choice_a": "one", + "choice_b": "fifteen", + "choice_c": "ground", + "choice_d": "no", + "answer_gt": "ground", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15294 + }, + { + "path": "Canada Geese Squawk on a Pond with a Fountain.wav", + "question": "What weather condition can be heard in the background?", + "choice_a": "rain", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15295 + }, + { + "path": "Silencyo_CC_Wind Mistral_Medium_Dry Grass_Crickets_Distant traffic.wav", + "question": "Is someone walking loudly?", + "choice_a": "never", + "choice_b": "seesaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15296 + }, + { + "path": "walking in gravel 2.wav", + "question": "Do the footsteps pause for a second?", + "choice_a": "train", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15297 + }, + { + "path": "WaterBottle.wav", + "question": "What noise does the man make?", + "choice_a": "tap", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "eating", + "answer_gt": "eating", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15298 + }, + { + "path": "Gulls at beach.wav", + "question": "Is this in nature?", + "choice_a": "five", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15299 + }, + { + "path": "April dawn chorus Sydenham Hill.wav", + "question": "Are the sounds coming from the forest?", + "choice_a": "drum", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15300 + }, + { + "path": "draining board metal drip on metal.wav", + "question": "Is the water running?", + "choice_a": "bird", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15301 + }, + { + "path": "creaky.wav", + "question": "Are there people listening to music?", + "choice_a": "clanking", + "choice_b": "sheep", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15302 + }, + { + "path": "mercury topaz starting.wav", + "question": "Does the car start?", + "choice_a": "crickets", + "choice_b": "medium", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15303 + }, + { + "path": "CAGE ELAVATOR MUMBAI.wav", + "question": "What is making the loud clanging noises?", + "choice_a": "door", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "door", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15304 + }, + { + "path": "Washing Machine Spins.wav", + "question": "Is a dog barking?", + "choice_a": "ground", + "choice_b": "person", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15305 + }, + { + "path": "Cable Car Ride Birds.wav", + "question": "Is something being opened?", + "choice_a": "four", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15306 + }, + { + "path": "Rain drops on marquee.wav", + "question": "is there thunder?", + "choice_a": "cards", + "choice_b": "bridge", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15307 + }, + { + "path": "Iceland2013_Stokkur.wav", + "question": "Are the people crying?", + "choice_a": "music", + "choice_b": "train", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15308 + }, + { + "path": "paper_cut.wav", + "question": "What is making the cutting sound?", + "choice_a": "five", + "choice_b": "chirp", + "choice_c": "yes", + "choice_d": "cutter", + "answer_gt": "cutter", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15309 + }, + { + "path": "helicopter.wav", + "question": "Is a motor making this sound?", + "choice_a": "people", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15310 + }, + { + "path": "squeaking wooden floor.wav", + "question": "Is this a sound that occurs naturally at the seaside?", + "choice_a": "tree", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15311 + }, + { + "path": "Rain falling on a metal roof - 96 kHz _ 24 Bit.wav", + "question": "What kind of object is the rain hitting?", + "choice_a": "snow", + "choice_b": "cleaning", + "choice_c": "yes", + "choice_d": "roof", + "answer_gt": "roof", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15312 + }, + { + "path": "Siren Milan.wav", + "question": "Is this a noise that could be heard in a city?", + "choice_a": "beeping", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15313 + }, + { + "path": "CAGE ELAVATOR MUMBAI.wav", + "question": "Does a door open?", + "choice_a": "twenty", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15314 + }, + { + "path": "12 noon church-bell 140310_0121.wav", + "question": "Are there birds?", + "choice_a": "road", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15315 + }, + { + "path": "draining board metal drip on metal.wav", + "question": "what is the water doing?", + "choice_a": "ground", + "choice_b": "motor", + "choice_c": "running", + "choice_d": "no", + "answer_gt": "running", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15316 + }, + { + "path": "Erny vs Deadman4.wav", + "question": "Are the people arguing in church?", + "choice_a": "crows", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15317 + }, + { + "path": "Walking alongside the road.wav", + "question": "How is the person moving?", + "choice_a": "fifteen", + "choice_b": "snow", + "choice_c": "walking", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15318 + }, + { + "path": "SilverStarSearchAndRescue.wav", + "question": "Is the sound constant?", + "choice_a": "thirty", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15319 + }, + { + "path": "coastal road on the beach, scooter, motorcycle, car.wav", + "question": "Did a helicopter take off?", + "choice_a": "three", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15320 + }, + { + "path": "box_open_hit.wav", + "question": "How many times does the sound repeat itself?", + "choice_a": "grass", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "nine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15321 + }, + { + "path": "laundry.machine.wav", + "question": "What is making the noise?", + "choice_a": "short", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "metal", + "answer_gt": "metal", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15322 + }, + { + "path": "crickets and owls.wav", + "question": "how many times is there a loud screeching sound?", + "choice_a": "hand", + "choice_b": "three", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15323 + }, + { + "path": "20080416.buzz.stereo.wav", + "question": "Did any object fall on the floor and break ?", + "choice_a": "yes", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15324 + }, + { + "path": "dripping.wav", + "question": "Is water babbling?", + "choice_a": "outside", + "choice_b": "mixer", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15325 + }, + { + "path": "Menziken Sawmill.wav", + "question": "Is the area crowded?", + "choice_a": "singing", + "choice_b": "party", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15326 + }, + { + "path": "FOLEY_Ext_Garbage_Hauling_001.wav", + "question": "Did it sound like there was something rolling around on a metal floor?", + "choice_a": "faucet", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15327 + }, + { + "path": "drip rhythm1.wav", + "question": "Is it thundering?", + "choice_a": "tearing", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15328 + }, + { + "path": "babbling brook 2.wav", + "question": "Is the liquid pouring?", + "choice_a": "ball", + "choice_b": "snow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15329 + }, + { + "path": "Room-tone rain-drips 1m 161015_1013.wav", + "question": "Are there animals around?", + "choice_a": "roof", + "choice_b": "fast", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15330 + }, + { + "path": "Wind moaning through gap in door and house noises.wav", + "question": "How is the weather?", + "choice_a": "flood", + "choice_b": "bike", + "choice_c": "windy", + "choice_d": "no", + "answer_gt": "windy", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15331 + }, + { + "path": "Flipping Coin Can.wav", + "question": "What is the size of the container used?", + "choice_a": "bubbling", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "small", + "answer_gt": "small", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15332 + }, + { + "path": "spring rain in the woods.wav", + "question": "Does the water dripping sound get louder?", + "choice_a": "bell", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15333 + }, + { + "path": "Centurion Suburb Evening.wav", + "question": "are any sounds of nature heard?", + "choice_a": "seven", + "choice_b": "barking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15334 + }, + { + "path": "paper_bag.wav", + "question": "What is being ripped?", + "choice_a": "short", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "paper", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15335 + }, + { + "path": "atmo_kenting_national_park.wav", + "question": "What does the animal heard here lay?", + "choice_a": "outside", + "choice_b": "owl", + "choice_c": "birds", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15336 + }, + { + "path": "Hanoi streets.wav", + "question": "What is heard beeping halfway through?", + "choice_a": "umbrella", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "car", + "answer_gt": "car", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15337 + }, + { + "path": "Rain on Car Roof 2.wav", + "question": "Does it sound like damage is being done to property?", + "choice_a": "static", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15338 + }, + { + "path": "diesel_train_comes.wav", + "question": "Are there multiple vehicles?", + "choice_a": "crows", + "choice_b": "twelve", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15339 + }, + { + "path": "SinkWater.wav", + "question": "Does water go down a drain?", + "choice_a": "cleaning", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15340 + }, + { + "path": "rotatingdome.wav", + "question": "Can a crowd be heard speaking in the background?", + "choice_a": "short", + "choice_b": "raining", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15341 + }, + { + "path": "small town.wav", + "question": "Is there a voice?", + "choice_a": "outdoors", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15342 + }, + { + "path": "More Amphitheatre Birds. Wav.wav", + "question": "What type of sounds are there?", + "choice_a": "sparrow", + "choice_b": "siren", + "choice_c": "yes", + "choice_d": "birds", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15343 + }, + { + "path": "Tools .wav", + "question": "what are the items doing with each other?", + "choice_a": "cat", + "choice_b": "clanking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "clanking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15344 + }, + { + "path": "ieai.wav", + "question": "is the water moving?", + "choice_a": "buzzsaw", + "choice_b": "twentyfive", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15345 + }, + { + "path": "13. Crushing tin can.wav", + "question": "How many milk cartons does this person open ?", + "choice_a": "paper", + "choice_b": "ocean", + "choice_c": "yes", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15346 + }, + { + "path": "Plaza_de_la_Revolucion_risa.wav", + "question": "What animal is barking?", + "choice_a": "steps", + "choice_b": "sink", + "choice_c": "yes", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15347 + }, + { + "path": "meadow brook bees.wav", + "question": "Is the machine whirring loudly?", + "choice_a": "engine", + "choice_b": "bear", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15348 + }, + { + "path": "cowshed.wav", + "question": "Is only one cow mooing?", + "choice_a": "outdoors", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15349 + }, + { + "path": "20060523.grassland.wav", + "question": "Is there a dog barking?", + "choice_a": "motorbike", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15350 + }, + { + "path": "second_floor_lav.wav", + "question": "Does the sound stay the same ?", + "choice_a": "rod", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15351 + }, + { + "path": "Walking_on_tarmac.wav", + "question": "Is the person near an ocean?", + "choice_a": "drink", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15352 + }, + { + "path": "street works_pressure_low rumble.wav", + "question": "What environment is this?", + "choice_a": "seven", + "choice_b": "outdoors", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "outdoors", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15353 + }, + { + "path": "river_mouth3.wav", + "question": "Is the area dry?", + "choice_a": "grass", + "choice_b": "sweeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15354 + }, + { + "path": "squeaking wooden floor.wav", + "question": "Does this sound involve objects that are being moved around?", + "choice_a": "singing", + "choice_b": "seven", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15355 + }, + { + "path": "Ship Fender.wav", + "question": "What are they using to start up the engine?", + "choice_a": "hands", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "start", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15356 + }, + { + "path": "auto-rickshaw-trip.wav", + "question": "Is there an animal?", + "choice_a": "rod", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15357 + }, + { + "path": "Kitchen faucet running fast and slow and filling glass of water.wav", + "question": "Is someone taking a bath?", + "choice_a": "umbrella", + "choice_b": "short", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15358 + }, + { + "path": "Index Card Flips _handle business paper mvmt_ 02.wav", + "question": "What is being shuffled?", + "choice_a": "crows", + "choice_b": "ball", + "choice_c": "cards", + "choice_d": "no", + "answer_gt": "cards", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15359 + }, + { + "path": "Wipers .wav", + "question": "Are the wipers of a car getting faster?", + "choice_a": "paper", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15360 + }, + { + "path": "spring, road.wav", + "question": "Is someone watering the plants ?", + "choice_a": "seven", + "choice_b": "umbrella", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15361 + }, + { + "path": "Parking Garage - Ambiance, Electrical Hum 1.wav", + "question": "Is a child crying about being hungry?", + "choice_a": "radio", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15362 + }, + { + "path": "glissando1b.wav", + "question": "Does the scraping sound continually overpowers the bell sound?", + "choice_a": "sound", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15363 + }, + { + "path": "fountains-Udaipur-Saheliyon-Ki-Bari-4.wav", + "question": "Are there no people nearby?", + "choice_a": "cutter", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15364 + }, + { + "path": "PIT-ROIG 0.12-0.17.wav", + "question": "Is there a dog barking?", + "choice_a": "road", + "choice_b": "stormy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15365 + }, + { + "path": "medical car horn EGYPT Alexandria.wav", + "question": "What is making the wa sound throughout the recording?", + "choice_a": "flood", + "choice_b": "men", + "choice_c": "siren", + "choice_d": "no", + "answer_gt": "siren", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15366 + }, + { + "path": "SPilling Water.wav", + "question": "Are there many people talking?", + "choice_a": "twentythree", + "choice_b": "announcement", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15367 + }, + { + "path": "windscreen wipers heavy rain.wav", + "question": "Is it raining?", + "choice_a": "canvas", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15368 + }, + { + "path": "Single cricket chirping during a summer evening in the city _with traffic noise_.wav", + "question": "what is making the high pitch sound?", + "choice_a": "voice", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "insect", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15369 + }, + { + "path": "Walking in Grass in Evening with Loud Bird.wav", + "question": "is the person walking on grass?", + "choice_a": "fifteen", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15370 + }, + { + "path": "BulletJuneEdited192012.wav", + "question": "What is making the sound?", + "choice_a": "saw", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "motorbike", + "answer_gt": "motorbike", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15371 + }, + { + "path": "channel 2 now concludes its broadcast day.wav", + "question": "Is this an electronic noise?", + "choice_a": "wrench", + "choice_b": "umbrella", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15372 + }, + { + "path": "08-Garage Opening-consolidated.wav", + "question": "Is the sound repitious?", + "choice_a": "screeching", + "choice_b": "static", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15373 + }, + { + "path": "Face slap CsG.wav", + "question": "How many times is the slapping sound heard?", + "choice_a": "road", + "choice_b": "rod", + "choice_c": "yes", + "choice_d": "twenty", + "answer_gt": "twenty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15374 + }, + { + "path": "UnionStation06OutBack_BusyOutside.wav", + "question": "Are people talking?", + "choice_a": "engine", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15375 + }, + { + "path": "17-Year Cicada Mating Call.wav", + "question": "What kind of noise is it?", + "choice_a": "siren", + "choice_b": "wood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "siren", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15376 + }, + { + "path": "Water Faucet HQ Stereo.wav", + "question": "Is water running?", + "choice_a": "washing", + "choice_b": "sink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15377 + }, + { + "path": "Trompetistas.wav", + "question": "are the people playing the wind instruments inside a room?", + "choice_a": "running", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15378 + }, + { + "path": "miniature goats and sheep.wav", + "question": "What animal is making the noise?", + "choice_a": "sheep", + "choice_b": "wood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15379 + }, + { + "path": "Squeaky Wood _Compilation_.wav", + "question": "What sound is the object making?", + "choice_a": "ball", + "choice_b": "screeching", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "screeching", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15380 + }, + { + "path": "Chime-Ball.wav", + "question": "Is the sound staticky?", + "choice_a": "honking", + "choice_b": "sound", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15381 + }, + { + "path": "Crickets in the night.wav", + "question": "What other animal is speaking?", + "choice_a": "nothing", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "nothing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15382 + }, + { + "path": "BoyRacer.wav", + "question": "Is someone driving?", + "choice_a": "wet", + "choice_b": "leaves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15383 + }, + { + "path": "Tree Bark Cracks.wav", + "question": "Is there a lot of people?", + "choice_a": "twenty", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15384 + }, + { + "path": "Grackles.wav", + "question": "Are there people talking?", + "choice_a": "one", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15385 + }, + { + "path": "squeaky metal swing.wav", + "question": "What playground feature makes this sound when in use?", + "choice_a": "ball", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "seesaw", + "answer_gt": "seesaw", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15386 + }, + { + "path": "RBH_Household_shower 03.wav", + "question": "Can a shower be heard?", + "choice_a": "whispering", + "choice_b": "intercom", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15387 + }, + { + "path": "h907 boules pologna clap f.wav", + "question": "Is someone addressing a crowd?", + "choice_a": "bridge", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15388 + }, + { + "path": "WaterOnMetal.wav", + "question": "What is making the noise?", + "choice_a": "brush", + "choice_b": "rain", + "choice_c": "machine", + "choice_d": "no", + "answer_gt": "machine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15389 + }, + { + "path": "Walking on dry grass.wav", + "question": "Are they walking on a busy street?", + "choice_a": "men", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15390 + }, + { + "path": "20060523.grassland.wav", + "question": "Is there wildlife close?", + "choice_a": "swimming", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15391 + }, + { + "path": "obresAranyo_trepant2.wav", + "question": "Is a tool being used?", + "choice_a": "sink", + "choice_b": "speaking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15392 + }, + { + "path": "Cars_bridge.wav", + "question": "What is the sound coming from?", + "choice_a": "traffic", + "choice_b": "people", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "traffic", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15393 + }, + { + "path": "mediterranean_sea_porticcio.wav", + "question": "Does the sound of the water alternate between soft and loud?", + "choice_a": "ten", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15394 + }, + { + "path": "Little sreet behind a terrasse cafe.wav", + "question": "Are people having a party here?", + "choice_a": "outdoors", + "choice_b": "men", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15395 + }, + { + "path": "channel 2 now concludes its broadcast day.wav", + "question": "What kind of building might have machinery that makes this noise?", + "choice_a": "thunder", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "beeping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15396 + }, + { + "path": "laundry.machine.wav", + "question": "What is making the noise?", + "choice_a": "wrench", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "wrench", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15397 + }, + { + "path": "Iceland2013_Stokkur.wav", + "question": "Are the people having a good time?", + "choice_a": "iron", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15398 + }, + { + "path": "Tiergarten birds early morning.wav", + "question": "what are the birds doing?", + "choice_a": "coin", + "choice_b": "grass", + "choice_c": "yes", + "choice_d": "singing", + "answer_gt": "singing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15399 + }, + { + "path": "10_lightning_kohchang.wav", + "question": "Is it snowing?", + "choice_a": "honking", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15400 + }, + { + "path": "Cityscape construction site 2 100304.wav", + "question": "Is this inside of a museum?", + "choice_a": "waves", + "choice_b": "canvas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15401 + }, + { + "path": "Water Faucet HQ Stereo.wav", + "question": "Are cats meowing?", + "choice_a": "hen", + "choice_b": "four", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15402 + }, + { + "path": "Cars_bridge.wav", + "question": "are horns honking?", + "choice_a": "tank", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15403 + }, + { + "path": "OrchestraTuning2.wav", + "question": "Is someone playing the harmonica?", + "choice_a": "start", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15404 + }, + { + "path": "INT Factory budimka, pozega.wav", + "question": "Does the flat slapping at the beginning happen more than once?", + "choice_a": "drill", + "choice_b": "cars", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15405 + }, + { + "path": "luffy_wind5.wav", + "question": "Is the sound consistent?", + "choice_a": "tearing", + "choice_b": "ten", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15406 + }, + { + "path": "steam engine at museum.wav", + "question": "Are there leaves being being blown by a blower?", + "choice_a": "tank", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15407 + }, + { + "path": "Bell_Hotel_desk.wav", + "question": "Does the bell stop and start?", + "choice_a": "water", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15408 + }, + { + "path": "PauseConference_youngerPeople.wav", + "question": "What are all the people in this area doing?", + "choice_a": "traffic", + "choice_b": "hand", + "choice_c": "speaking", + "choice_d": "no", + "answer_gt": "speaking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15409 + }, + { + "path": "Rave1.wav", + "question": "It the sound playing rhythmically proportional?", + "choice_a": "stops", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15410 + }, + { + "path": "Page turns and book close_open.wav", + "question": "Is it a repetitive sound?", + "choice_a": "birds", + "choice_b": "mouse", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15411 + }, + { + "path": "Blackbird tweet with waterfall background.wav", + "question": "What is the water doing?", + "choice_a": "falling", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "falling", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15412 + }, + { + "path": "at the edge of the forest.wav", + "question": "Is there wind?", + "choice_a": "train", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15413 + }, + { + "path": "squeaking wooden floor.wav", + "question": "Is this a sound that occurs naturally at the seaside?", + "choice_a": "slowly", + "choice_b": "wood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15414 + }, + { + "path": "cricket chirp.wav", + "question": "Is it night time?", + "choice_a": "one", + "choice_b": "iron", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15415 + }, + { + "path": "Scratching or tearing.wav", + "question": "is there a scraping sound?", + "choice_a": "small", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15416 + }, + { + "path": "living room tone ambient distant noises neighbours.wav", + "question": "are the machines working loudly?", + "choice_a": "bus", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15417 + }, + { + "path": "Babble of Frogs 001.wav", + "question": "Is this a virtual location?", + "choice_a": "bell", + "choice_b": "glass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15418 + }, + { + "path": "005_musesdelight_charismatic-african-preacher.wav", + "question": "Is this a place to go shopping?", + "choice_a": "clanking", + "choice_b": "never", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15419 + }, + { + "path": "Glass Chandelier Rattle.wav", + "question": "What type sound wave is being produced?", + "choice_a": "falls", + "choice_b": "loudspeaker", + "choice_c": "birds", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15420 + }, + { + "path": "Sukhapha anchor chain.wav", + "question": "what does the person rattle?", + "choice_a": "engine", + "choice_b": "wind", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "engine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15421 + }, + { + "path": "Fowl - Chatter 1 - 96kHz.wav", + "question": "What animal is making the most noise?", + "choice_a": "barking", + "choice_b": "chicken", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15422 + }, + { + "path": "Car Engine Idling.wav", + "question": "What other sound was heard?", + "choice_a": "mixer", + "choice_b": "intercom", + "choice_c": "yes", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15423 + }, + { + "path": "20061124ParadeCHS.wav", + "question": "Where is the music coming from?", + "choice_a": "band", + "choice_b": "twenty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "band", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15424 + }, + { + "path": "train.wav", + "question": "Is there a whistling noise?", + "choice_a": "windy", + "choice_b": "constant", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15425 + }, + { + "path": "SCC CLAPTER 20101210.wav", + "question": "Are people cheering?", + "choice_a": "static", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15426 + }, + { + "path": "Watering Can.wav", + "question": "Is gravel being poured?", + "choice_a": "once", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15427 + }, + { + "path": "April dawn chorus Sydenham Hill.wav", + "question": "Are the sounds coming from the forest?", + "choice_a": "cabinet", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15428 + }, + { + "path": "Fuente Cotino 2.wav", + "question": "Is someone urinating?", + "choice_a": "party", + "choice_b": "short", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15429 + }, + { + "path": "Saas-Fee Village Atmosphere and Church 100611.wav", + "question": "How many times does the bell ring?", + "choice_a": "birds", + "choice_b": "eight", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "eight", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15430 + }, + { + "path": "outdoors ambient distant village 3.wav", + "question": "Is there a car horn sound at any point?", + "choice_a": "twentyfive", + "choice_b": "traffic", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15431 + }, + { + "path": "ortam.wav", + "question": "When is the pitch of the rumbling sound the highest?", + "choice_a": "cow", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "constant", + "answer_gt": "constant", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15432 + }, + { + "path": "water_stream_001.wav", + "question": "Where is the water coming from?", + "choice_a": "wet", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "faucet", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15433 + }, + { + "path": "Ambience with Train.wav", + "question": "Is a horn being honked?", + "choice_a": "hall", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15434 + }, + { + "path": "California morning birds singing.wav", + "question": "What breed of bird is chirping?", + "choice_a": "rod", + "choice_b": "machine", + "choice_c": "sparrow", + "choice_d": "no", + "answer_gt": "sparrow", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15435 + }, + { + "path": "small town.wav", + "question": "What is humming?", + "choice_a": "dog", + "choice_b": "cleaning", + "choice_c": "vehicle", + "choice_d": "no", + "answer_gt": "vehicle", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15436 + }, + { + "path": "thunderstorm_2.wav", + "question": "Is there a rainstorm?", + "choice_a": "motorbike", + "choice_b": "seesaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15437 + }, + { + "path": "Kung Fu Clothes Hits and Clothing Sounds.wav", + "question": "how many taps are heard?", + "choice_a": "chain", + "choice_b": "ten", + "choice_c": "yes", + "choice_d": "eight", + "answer_gt": "eight", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15438 + }, + { + "path": "Lambs.wav", + "question": "Is this outdoors?", + "choice_a": "waves", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15439 + }, + { + "path": "carkeys.wav", + "question": "What is the person holding?", + "choice_a": "vehicle", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "chain", + "answer_gt": "chain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15440 + }, + { + "path": "faucet3.wav", + "question": "Is the appliance making the noise at full power in the middle of the clip?", + "choice_a": "wet", + "choice_b": "normal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15441 + }, + { + "path": "SpringPeepersMarch2012.wav", + "question": "What word describes the main sound in this clip?", + "choice_a": "two", + "choice_b": "crickets", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "crickets", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15442 + }, + { + "path": "London Overground train _interior_ approaches Victoria Station.wav", + "question": "Are the people currently in transit?", + "choice_a": "intercom", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15443 + }, + { + "path": "Fuente Cotino 2.wav", + "question": "Is the water hitting a man-made structure?", + "choice_a": "sweeping", + "choice_b": "mouse", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15444 + }, + { + "path": "Calle.wav", + "question": "Where are they?", + "choice_a": "liquid", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "road", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15445 + }, + { + "path": "WavesOnTheShore.wav", + "question": "What is the person moving in?", + "choice_a": "metal", + "choice_b": "bird", + "choice_c": "swimming", + "choice_d": "no", + "answer_gt": "swimming", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15446 + }, + { + "path": "Waterfall close.wav", + "question": "Is the sound at a steady volume?", + "choice_a": "crows", + "choice_b": "cards", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15447 + }, + { + "path": "river_mouth1.wav", + "question": "Is the air rushing sound a vehicle of any type?", + "choice_a": "party", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15448 + }, + { + "path": "InsectsSummerSwellingAug16th2013.wav", + "question": "What metal is being carved?", + "choice_a": "start", + "choice_b": "rod", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rod", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15449 + }, + { + "path": "Krankenwagen _ German Ambulances Passing by...wav", + "question": "Where is the siren coming from?", + "choice_a": "bike", + "choice_b": "snow", + "choice_c": "yes", + "choice_d": "police", + "answer_gt": "police", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15450 + }, + { + "path": "indoors house dorm dormitory ambient room tone distant motor.wav", + "question": "Is an engine making a sound?", + "choice_a": "distortion", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15451 + }, + { + "path": "fresound sample 2.wav", + "question": "How many times does the note change?", + "choice_a": "five", + "choice_b": "wind", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15452 + }, + { + "path": "Ambience - Cattle Barn - Busy - 96kHzhg.wav", + "question": "How many types of animals can be heard making noise ?", + "choice_a": "motorbike", + "choice_b": "mixer", + "choice_c": "three", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15453 + }, + { + "path": "St. Marks NWR Wind at the Marina.wav", + "question": "Is it a wet day?", + "choice_a": "four", + "choice_b": "six", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15454 + }, + { + "path": "abandoned-ballroom-radiators.wav", + "question": "Is the sound melodic?", + "choice_a": "radio", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15455 + }, + { + "path": "Bicycle Chain Accel Crash.wav", + "question": "Is the sound being made constant?", + "choice_a": "brush", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15456 + }, + { + "path": "Machete vs frying pan 2.wav", + "question": "Does the sound stop in the middle?", + "choice_a": "wood", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15457 + }, + { + "path": "110724_inriversidemus1.wav", + "question": "Are there any animals making sounds?", + "choice_a": "two", + "choice_b": "men", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15458 + }, + { + "path": "incheon_4am.wav", + "question": "Are there voices only at the beginning?", + "choice_a": "saw", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15459 + }, + { + "path": "Car Engine Idling.wav", + "question": "What kind of motor is running?", + "choice_a": "interested", + "choice_b": "distortion", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "interested", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15460 + }, + { + "path": "Outside wind.wav", + "question": "Is the chair squeaking?", + "choice_a": "steps", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15461 + }, + { + "path": "Metal_On_Wood_Hits_Axe.wav", + "question": "What activity is taking place?", + "choice_a": "hands", + "choice_b": "chopping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "chopping", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15462 + }, + { + "path": "Growing Hum.wav", + "question": "How many times is the clicking heard at the beginning?", + "choice_a": "six", + "choice_b": "seesaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "six", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15463 + }, + { + "path": "urinating on a wall.wav", + "question": "what is coming out of the hose?", + "choice_a": "electronic", + "choice_b": "bell", + "choice_c": "water", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15464 + }, + { + "path": "T24_home_dishwasher_cycle change.wav", + "question": "Is machinery involved?", + "choice_a": "thunder", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15465 + }, + { + "path": "water slushing moderate speed.wav", + "question": "What liquid is being played with?", + "choice_a": "twice", + "choice_b": "bell", + "choice_c": "water", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15466 + }, + { + "path": "Small growling dog.wav", + "question": "Is there someone laughing?", + "choice_a": "ocean", + "choice_b": "running", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15467 + }, + { + "path": "Birds-Crow & Song Birds.wav", + "question": "Are these noises coming from outdoors ?", + "choice_a": "six", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15468 + }, + { + "path": "Evening Atmosphere #2.wav", + "question": "How many times does the dog bark?", + "choice_a": "nine", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "nine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15469 + }, + { + "path": "small dog leaves.wav", + "question": "Is there a crumbling noise?", + "choice_a": "no", + "choice_b": "cricket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15470 + }, + { + "path": "0221 Bar_terrace.wav", + "question": "Are there only males?", + "choice_a": "seagulls", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15471 + }, + { + "path": "radio tuning 2.wav", + "question": "What electronic is being dialed?", + "choice_a": "big", + "choice_b": "paper", + "choice_c": "yes", + "choice_d": "radio", + "answer_gt": "radio", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15472 + }, + { + "path": "filling-ice-cup.wav", + "question": "What can the person do with the liquid next?", + "choice_a": "train", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "drink", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15473 + }, + { + "path": "windy rain.wav", + "question": "What is making the constant sound?", + "choice_a": "paper", + "choice_b": "music", + "choice_c": "rain", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15474 + }, + { + "path": "Wind-up Crank.wav", + "question": "Is the squealing noise unnatural?", + "choice_a": "nothing", + "choice_b": "heavy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15475 + }, + { + "path": "Muddy_steps_bush_birds_singing.wav", + "question": "Is this person inside of a house?", + "choice_a": "hands", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15476 + }, + { + "path": "rain in tent.wav", + "question": "what is the rain doing?", + "choice_a": "car", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "raining", + "answer_gt": "raining", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15477 + }, + { + "path": "earth_movement.wav", + "question": "This sound is reminiscent of what domesticated animal when it's happy?", + "choice_a": "sand", + "choice_b": "cat", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15478 + }, + { + "path": "Small growling dog.wav", + "question": "Is a dog growling ?", + "choice_a": "seesaw", + "choice_b": "barking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15479 + }, + { + "path": "fireTruckFar NL 140109_00.wav", + "question": "Did a cup fall to the ground and break ?", + "choice_a": "singing", + "choice_b": "metal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15480 + }, + { + "path": "024_House_InsideCarEngineStart.wav", + "question": "Is the engine off at the beginning?", + "choice_a": "distortion", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15481 + }, + { + "path": "spring rain in the woods.wav", + "question": "Does the water stop dripping?", + "choice_a": "locust", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15482 + }, + { + "path": "Balloon Game at Arlington Heights Carnival.wav", + "question": "What kind of toy is the child asking for?", + "choice_a": "sparrow", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15483 + }, + { + "path": "Running.wav", + "question": "does the person walk then run?", + "choice_a": "owl", + "choice_b": "medium", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15484 + }, + { + "path": "13. Crushing tin can.wav", + "question": "How many milk cartons does this person open ?", + "choice_a": "raining", + "choice_b": "grass", + "choice_c": "two", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15485 + }, + { + "path": "Atmosfera Miasto Spokojna dzielnica rano.wav", + "question": "is there more than one car?", + "choice_a": "nothing", + "choice_b": "small", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15486 + }, + { + "path": "fountain.wav", + "question": "Is it raining?", + "choice_a": "men", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15487 + }, + { + "path": "spring, road.wav", + "question": "What type of material is making that impact noise ?", + "choice_a": "roof", + "choice_b": "buzzing", + "choice_c": "wood", + "choice_d": "no", + "answer_gt": "wood", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15488 + }, + { + "path": "Small Boat Engine.wav", + "question": "Is this a person?", + "choice_a": "wet", + "choice_b": "bike", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15489 + }, + { + "path": "Strong wind in trees.wav", + "question": "What machine is heard?", + "choice_a": "static", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "engine", + "answer_gt": "engine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15490 + }, + { + "path": "treefrogs.wav", + "question": "Is the hissing decreasing in volume over time?", + "choice_a": "tank", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15491 + }, + { + "path": "crickets and owls.wav", + "question": "is there a lot of traffic noise?", + "choice_a": "glass", + "choice_b": "radio", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15492 + }, + { + "path": "Charleston Campus Summer.wav", + "question": "Can any cars be heard driving by?", + "choice_a": "cow", + "choice_b": "tank", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15493 + }, + { + "path": "Bus Pulls Away.wav", + "question": "Is the car driving away?", + "choice_a": "driving", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15494 + }, + { + "path": "InsectsSummerSwellingAug16th2013.wav", + "question": "What metal is being carved?", + "choice_a": "drill", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "iron", + "answer_gt": "iron", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15495 + }, + { + "path": "amazon 04.wav", + "question": "What is the predominant animal heard?", + "choice_a": "big", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15496 + }, + { + "path": "Toilet Shuffling.wav", + "question": "Is someone speaking?", + "choice_a": "honking", + "choice_b": "clanking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15497 + }, + { + "path": "thunderstorm_2.wav", + "question": "Is there thunder during the rain?", + "choice_a": "seagulls", + "choice_b": "tearing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15498 + }, + { + "path": "Walking in Grass in Evening with Loud Bird.wav", + "question": "Could the person be in walking in a room?", + "choice_a": "hen", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15499 + }, + { + "path": "Door Slam.wav", + "question": "Is an animal making noise?", + "choice_a": "small", + "choice_b": "sink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15500 + }, + { + "path": "Serving Water Quickly.wav", + "question": "Are several glasses being poured?", + "choice_a": "honking", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15501 + }, + { + "path": "WindInPylons.wav", + "question": "Is this inside a train ?", + "choice_a": "talking", + "choice_b": "glass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15502 + }, + { + "path": "Drill press.wav", + "question": "What is put in a combustible engine to make it run?", + "choice_a": "twice", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "gas", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15503 + }, + { + "path": "Fairground 2 Ghost ride.wav", + "question": "is the area a family friendly environment?", + "choice_a": "tree", + "choice_b": "bucket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15504 + }, + { + "path": "abandoned-ballroom-big-metal.wav", + "question": "What is crashing?", + "choice_a": "lock", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "tank", + "answer_gt": "tank", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15505 + }, + { + "path": "rainy stream 22 sec.wav", + "question": "Is it raining?", + "choice_a": "sand", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15506 + }, + { + "path": "upstairs.wav", + "question": "Is music played in the background ?", + "choice_a": "rain", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15507 + }, + { + "path": "sizzling oil.wav", + "question": "How many taps are there?", + "choice_a": "water", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15508 + }, + { + "path": "greece_melanes_cofee_1.wav", + "question": "What interrupts the person talking on the phone?", + "choice_a": "brush", + "choice_b": "tree", + "choice_c": "yes", + "choice_d": "person", + "answer_gt": "person", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15509 + }, + { + "path": "Gentle rain outside balcony street noise.wav", + "question": "How many engines can be heard?", + "choice_a": "ten", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15510 + }, + { + "path": "Vomit, puking spilling water onto grass splat.wav", + "question": "Is something being poured out?", + "choice_a": "insect", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15511 + }, + { + "path": "alarm_clock_ringing_01.wav", + "question": "does the ringing stop?", + "choice_a": "women", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15512 + }, + { + "path": "Neighbourhood evening ambience.wav", + "question": "Is this outside?", + "choice_a": "engine", + "choice_b": "sound", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15513 + }, + { + "path": "Lambs.wav", + "question": "Does the chirping become noticeably louder?", + "choice_a": "metal", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15514 + }, + { + "path": "indoors dorm dormitory ambient room tone distant traffic in street.wav", + "question": "What is honking?", + "choice_a": "car", + "choice_b": "fireworks", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "car", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15515 + }, + { + "path": "metal-bell-percussion.wav", + "question": "how many pauses are between each set of ringing?", + "choice_a": "twentythree", + "choice_b": "hand", + "choice_c": "two", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15516 + }, + { + "path": "Gulls at beach.wav", + "question": "How many people can be heard talking?", + "choice_a": "birds", + "choice_b": "chain", + "choice_c": "none", + "choice_d": "no", + "answer_gt": "none", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15517 + }, + { + "path": "Whalesong.wav", + "question": "What is this aniimal?", + "choice_a": "iron", + "choice_b": "ambulance", + "choice_c": "bear", + "choice_d": "no", + "answer_gt": "bear", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15518 + }, + { + "path": "SCC CLAPTER 20101210.wav", + "question": "Is there only a single person in the audience?", + "choice_a": "twentythree", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15519 + }, + { + "path": "vague_sable.wav", + "question": "What is it water called when it builds up and crashes on the sand?", + "choice_a": "chopping", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "waves", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15520 + }, + { + "path": "Mulholland Memorial Fountain Los Angeles.wav", + "question": "Is this indoors?", + "choice_a": "thunder", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15521 + }, + { + "path": "Night in nature.wav", + "question": "Does the noise get softer?", + "choice_a": "gas", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15522 + }, + { + "path": "glas-bubbels-def01.wav", + "question": "what is the water doing?", + "choice_a": "iron", + "choice_b": "bird", + "choice_c": "bubbling", + "choice_d": "no", + "answer_gt": "bubbling", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15523 + }, + { + "path": "INT Factory budimka, pozega.wav", + "question": "Are birds chirping in the background?", + "choice_a": "seven", + "choice_b": "constant", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15524 + }, + { + "path": "Steam 20.wav", + "question": "Does a soda can open?", + "choice_a": "ocean", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15525 + }, + { + "path": "walkingondirtpath.wav", + "question": "Is the man walking with anyone?", + "choice_a": "no", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15526 + }, + { + "path": "Beach_SaintJeanDeLuz_France_Waves_Kids_People_xystereo.wav", + "question": "Are there any kids playing?", + "choice_a": "seagulls", + "choice_b": "outdoors", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15527 + }, + { + "path": "Kitchen Noise From Distance.wav", + "question": "Are these noises something one might hear in a kitchen?", + "choice_a": "faucet", + "choice_b": "tank", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15528 + }, + { + "path": "RBH_Household_shower 03.wav", + "question": "Can a shower be heard?", + "choice_a": "windy", + "choice_b": "drilling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15529 + }, + { + "path": "Footsteps Dress Shoes Wood Floor.wav", + "question": "Is this person wearing shoes?", + "choice_a": "talking", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15530 + }, + { + "path": "cowshed.wav", + "question": "What animal is making the noise?", + "choice_a": "cow", + "choice_b": "bubbling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "cow", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15531 + }, + { + "path": "pencil on paper.wav", + "question": "What is someone writing on?", + "choice_a": "drum", + "choice_b": "paper", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15532 + }, + { + "path": "water fountian-hollow sm med pitch.wav", + "question": "What is the person doing in the bathroom?", + "choice_a": "big", + "choice_b": "ten", + "choice_c": "showering", + "choice_d": "no", + "answer_gt": "showering", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15533 + }, + { + "path": "01 residential elevator with door open and close.wav", + "question": "Is there water dripping?", + "choice_a": "grass", + "choice_b": "eating", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15534 + }, + { + "path": "Office Lift 2.wav", + "question": "What is opened?", + "choice_a": "ambulance", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "door", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15535 + }, + { + "path": "carnival_parade_cologne_1.wav", + "question": "Is this an adults only crowd?", + "choice_a": "bird", + "choice_b": "thunder", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15536 + }, + { + "path": "dragged-glass-object.wav", + "question": "Is the thing doing the scratching moving in different directions?", + "choice_a": "lightning", + "choice_b": "ambulance", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15537 + }, + { + "path": "Ship Fender.wav", + "question": "Is the man driving a bus?", + "choice_a": "seesaw", + "choice_b": "men", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15538 + }, + { + "path": "cricket chirp.wav", + "question": "Is the sound made by an insect?", + "choice_a": "machine", + "choice_b": "bike", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15539 + }, + { + "path": "So_little_time_so_many_frequencies.wav", + "question": "Does this involve sound effects?", + "choice_a": "bucket", + "choice_b": "mouse", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15540 + }, + { + "path": "Glass bottles in and out of a basket.wav", + "question": "How many objects are dragged?", + "choice_a": "seven", + "choice_b": "loudspeaker", + "choice_c": "glass", + "choice_d": "no", + "answer_gt": "glass", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15541 + }, + { + "path": "outdoors ambient village bird distant neighbours children car.wav", + "question": "What else can be heard?", + "choice_a": "birds", + "choice_b": "raining", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15542 + }, + { + "path": "10_lightning_kohchang.wav", + "question": "What is happening that causes this sound?", + "choice_a": "slowly", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "lightning", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15543 + }, + { + "path": "Cooking rice.wav", + "question": "What weather is happening outside?", + "choice_a": "buzzsaw", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15544 + }, + { + "path": "Ambience with Train.wav", + "question": "Is a party happening?", + "choice_a": "twenty", + "choice_b": "barking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15545 + }, + { + "path": "Underwater_rumble_CsG.wav", + "question": "Does the rumbling ever stop?", + "choice_a": "train", + "choice_b": "ambulance", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15546 + }, + { + "path": "Grackles.wav", + "question": "what type of building where people pay to see animals could this be located in?", + "choice_a": "zoo", + "choice_b": "low", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "zoo", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15547 + }, + { + "path": "Oppedette cafe #1.wav", + "question": "Does a man sing at the end?", + "choice_a": "static", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15548 + }, + { + "path": "indoors dorm dormitory ambient room tone distant traffic in street.wav", + "question": "Are there people running?", + "choice_a": "ambulance", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15549 + }, + { + "path": "Room-tone rain-drips 1m 161015_1013.wav", + "question": "Are the middle and final parts of the sound rhythmic?", + "choice_a": "three", + "choice_b": "chirping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15550 + }, + { + "path": "room-tone theater with silent woman 130525_07.wav", + "question": "does a person speak?", + "choice_a": "dog", + "choice_b": "forty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15551 + }, + { + "path": "espresso-maschine.wav", + "question": "What machine is making the loud grumbling noise?", + "choice_a": "mixer", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "mixer", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15552 + }, + { + "path": "Fast stream _ small river.wav", + "question": "What liquid is making the rushing noise?", + "choice_a": "water", + "choice_b": "car", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15553 + }, + { + "path": "fan_2_300513.wav", + "question": "What is blowing by?", + "choice_a": "fireworks", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "traffic", + "answer_gt": "traffic", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15554 + }, + { + "path": "wind and birds in the delta of the River Po 2.wav", + "question": "Are people talking?", + "choice_a": "buzzing", + "choice_b": "chicken", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15555 + }, + { + "path": "Water Faucet HQ Stereo.wav", + "question": "What moving substance is causing this sound?", + "choice_a": "storm", + "choice_b": "chirping", + "choice_c": "water", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15556 + }, + { + "path": "Lots of Geese.wav", + "question": "Are geese making noise?", + "choice_a": "flood", + "choice_b": "seagulls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15557 + }, + { + "path": "INT Factory budimka, pozega.wav", + "question": "Are birds chirping in the background?", + "choice_a": "cards", + "choice_b": "cutter", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15558 + }, + { + "path": "Organic sound.wav", + "question": "Is water running?", + "choice_a": "low", + "choice_b": "men", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15559 + }, + { + "path": "Construction Sounds.wav", + "question": "Is this loud?", + "choice_a": "building", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15560 + }, + { + "path": "People talking while waiting the bus.wav", + "question": "Are the people singing together?", + "choice_a": "seesaw", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15561 + }, + { + "path": "By ther blacksmith-002.wav", + "question": "Is everyone silent in the area?", + "choice_a": "singing", + "choice_b": "hand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15562 + }, + { + "path": "md3trk2.wav", + "question": "Is the noise repetitious?", + "choice_a": "wood", + "choice_b": "showering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15563 + }, + { + "path": "STE-031 italian road.wav", + "question": "Is the hosepipe running constantly?", + "choice_a": "singing", + "choice_b": "crows", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15564 + }, + { + "path": "005_musesdelight_charismatic-african-preacher.wav", + "question": "Is there a bell ringing?", + "choice_a": "nine", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15565 + }, + { + "path": "17-Year Cicada Mating Call.wav", + "question": "How many times does the alarm go off ?", + "choice_a": "six", + "choice_b": "medium", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "six", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15566 + }, + { + "path": "OiseauNuit1.wav", + "question": "Is the source of the chirping up above?", + "choice_a": "traffic", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15567 + }, + { + "path": "remix of 130879__frederic-font__05-hang-song-1.wav", + "question": "Is it repetitive?", + "choice_a": "cat", + "choice_b": "drum", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15568 + }, + { + "path": "BulletJuneEdited192012.wav", + "question": "Does this sound get quieter as the clip goes on?", + "choice_a": "working", + "choice_b": "whispering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15569 + }, + { + "path": "Dog escapes from the room.wav", + "question": "What slams closed?", + "choice_a": "door", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "door", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15570 + }, + { + "path": "20160506_sharpening.02.wav", + "question": "Does the sound stay at the same volume?", + "choice_a": "five", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15571 + }, + { + "path": "Small growling dog.wav", + "question": "Are people singing ?", + "choice_a": "one", + "choice_b": "bucket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15572 + }, + { + "path": "F907 Church prayer f.wav", + "question": "How many women speak?", + "choice_a": "person", + "choice_b": "cricket", + "choice_c": "yes", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15573 + }, + { + "path": "Pigeon Temple.wav", + "question": "Is a band playing music?", + "choice_a": "train", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15574 + }, + { + "path": "Trompetistas.wav", + "question": "are the people playing the wind instruments inside a room?", + "choice_a": "announcement", + "choice_b": "clapping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15575 + }, + { + "path": "drunk_teenagers_1.wav", + "question": "Are birds chirping?", + "choice_a": "gas", + "choice_b": "motor", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15576 + }, + { + "path": "windup_dino_slow.wav", + "question": "Is there a click clack sound throughout?", + "choice_a": "bridge", + "choice_b": "speaking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15577 + }, + { + "path": "Night Frogs.wav", + "question": "Is this noise consistent?", + "choice_a": "engine", + "choice_b": "clanging", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15578 + }, + { + "path": "Rushing_water+wind-Rec_Samsung_HMX-F80_Camcorder.wav", + "question": "Can boats be heard?", + "choice_a": "walking", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15579 + }, + { + "path": "obresAranyo_trepant2.wav", + "question": "What is the tool that the person is operating?", + "choice_a": "buzzsaw", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "buzzsaw", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15580 + }, + { + "path": "pr#6F9A9E.wav", + "question": "What is the object being dropped?", + "choice_a": "twentythree", + "choice_b": "wrench", + "choice_c": "rod", + "choice_d": "no", + "answer_gt": "rod", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15581 + }, + { + "path": "AMB_COLE.wav", + "question": "Is this a small gathering?", + "choice_a": "sound", + "choice_b": "showering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15582 + }, + { + "path": "Birds-Crow & Song Birds.wav", + "question": "Is a car honking ?", + "choice_a": "birds", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15583 + }, + { + "path": "Bathtub_Water-drain.wav", + "question": "What is the persons location to create this sound?", + "choice_a": "tap", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "tap", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15584 + }, + { + "path": "Metal_On_Wood_Hits_Axe.wav", + "question": "Is the sound interrupted at any point?", + "choice_a": "drill", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15585 + }, + { + "path": "steam engine at museum.wav", + "question": "Is the machine keeping the same rhythm?", + "choice_a": "radio", + "choice_b": "grass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15586 + }, + { + "path": "3. Ambience interior car windy.wav", + "question": "Are people talking?", + "choice_a": "driving", + "choice_b": "tank", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15587 + }, + { + "path": "bierfest_atmosphere.wav", + "question": "Is there barking?", + "choice_a": "roof", + "choice_b": "waves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15588 + }, + { + "path": "Cable Car Ride Birds.wav", + "question": "Is something being opened?", + "choice_a": "grass", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15589 + }, + { + "path": "birds chirping 03 short.wav", + "question": "What sound does a bird make?", + "choice_a": "constant", + "choice_b": "electronic", + "choice_c": "chirp", + "choice_d": "no", + "answer_gt": "chirp", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15590 + }, + { + "path": "Marker Writing on Paper.wav", + "question": "Is the person writing fast?", + "choice_a": "men", + "choice_b": "sheep", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15591 + }, + { + "path": "quedlinburg castle.wav", + "question": "What kind of mouth does the animal making sounds have?", + "choice_a": "cars", + "choice_b": "insect", + "choice_c": "owl", + "choice_d": "no", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15592 + }, + { + "path": "wind-sound-from-inside-car.wav", + "question": "Is it windy?", + "choice_a": "thirty", + "choice_b": "open", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15593 + }, + { + "path": "Ext-amb_park_late-fall_distant-gun-shot_Distant-Child-shouting.wav", + "question": "Is a parrot squawking?", + "choice_a": "motorbike", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15594 + }, + { + "path": "outdoors ambient windy wind leaves rustle hum.wav", + "question": "What is falling down?", + "choice_a": "never", + "choice_b": "leaves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "leaves", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15595 + }, + { + "path": "Curtain.wav", + "question": "Is someone dragging something?", + "choice_a": "music", + "choice_b": "low", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15596 + }, + { + "path": "urinating on a wall.wav", + "question": "How many water taps are open ?", + "choice_a": "band", + "choice_b": "twentythree", + "choice_c": "yes", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15597 + }, + { + "path": "Thunder Outside.wav", + "question": "How many strikes of thunder are there?", + "choice_a": "rainy", + "choice_b": "whispering", + "choice_c": "two", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15598 + }, + { + "path": "Glass bottles in and out of a basket.wav", + "question": "Does any glass break?", + "choice_a": "nothing", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15599 + }, + { + "path": "at the edge of the forest.wav", + "question": "Is there wind?", + "choice_a": "saw", + "choice_b": "sound", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15600 + }, + { + "path": "Cityscape construction site 2 100304.wav", + "question": "Can a car be heard in the background?", + "choice_a": "outdoors", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15601 + }, + { + "path": "birdy.wav", + "question": "is there an airplane flying by?", + "choice_a": "small", + "choice_b": "tank", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15602 + }, + { + "path": "Face slap CsG.wav", + "question": "Is there a slapping sound?", + "choice_a": "barking", + "choice_b": "twice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15603 + }, + { + "path": "CAR_WASH.wav", + "question": "What is an example of something carried by this vehicle?", + "choice_a": "people", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "people", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15604 + }, + { + "path": "Walla chatter, adults and children in auditorium.wav", + "question": "Is there more than one person in the group?", + "choice_a": "wet", + "choice_b": "flushing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15605 + }, + { + "path": "LightRaininPinesMarch302013.wav", + "question": "What is the rain hitting?", + "choice_a": "driving", + "choice_b": "window", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "window", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15606 + }, + { + "path": "Running.wav", + "question": "Is the person driving?", + "choice_a": "medium", + "choice_b": "flushing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15607 + }, + { + "path": "branch and wind in wood 1.wav", + "question": "Are there people talking?", + "choice_a": "ocean", + "choice_b": "small", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15608 + }, + { + "path": "Fuente Cotino 2.wav", + "question": "Is someone urinating?", + "choice_a": "beeping", + "choice_b": "ground", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15609 + }, + { + "path": "Train sound.wav", + "question": "is something moving?", + "choice_a": "people", + "choice_b": "seagulls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15610 + }, + { + "path": "Thunder 03.wav", + "question": "It is the sound of an inanimate object?", + "choice_a": "leaves", + "choice_b": "thirty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15611 + }, + { + "path": "steps_snow.wav", + "question": "What does the group keep doing?", + "choice_a": "radio", + "choice_b": "walking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15612 + }, + { + "path": "20080416.bunting.wav", + "question": "are there people yelling?", + "choice_a": "nine", + "choice_b": "tap", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15613 + }, + { + "path": "walking in gravel 2.wav", + "question": "Is the person indoors?", + "choice_a": "washing", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15614 + }, + { + "path": "Steam 20.wav", + "question": "What instrument is being played?", + "choice_a": "drum", + "choice_b": "fast", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "drum", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15615 + }, + { + "path": "Swings in Mauerpark, Berlin.wav", + "question": "Does the object continue to make noise?", + "choice_a": "bridge", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15616 + }, + { + "path": "Still morning Collingwood beach.wav", + "question": "Is a motorboat moving through the water?", + "choice_a": "voice", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15617 + }, + { + "path": "WalkingInSnowCrunchingIce.wav", + "question": "Is a cricket chirping?", + "choice_a": "ground", + "choice_b": "seven", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15618 + }, + { + "path": "20080918.boots.door.wav", + "question": "Is there a person walking?", + "choice_a": "announcement", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15619 + }, + { + "path": "SonicSnap_GPSUK_Cockerel.wav", + "question": "What bird is making a sound near the end?", + "choice_a": "rooster", + "choice_b": "loudspeaker", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rooster", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15620 + }, + { + "path": "Lots of Geese.wav", + "question": "Are people yelling?", + "choice_a": "building", + "choice_b": "twentyfive", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15621 + }, + { + "path": "OiseauNuit1.wav", + "question": "What is singing?", + "choice_a": "birds", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "birds", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15622 + }, + { + "path": "Walking alongside the road.wav", + "question": "Do they open a door?", + "choice_a": "gas", + "choice_b": "clanging", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15623 + }, + { + "path": "Elbe near Ovelgoenne.wav", + "question": "Is the person taking a shower?", + "choice_a": "brush", + "choice_b": "rooster", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15624 + }, + { + "path": "worktoilet.wav", + "question": "Is there animal sounds?", + "choice_a": "brush", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15625 + }, + { + "path": "knocking on a window or glass.wav", + "question": "Is this an animal?", + "choice_a": "small", + "choice_b": "coin", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15626 + }, + { + "path": "WasherEndofRestCycleStartFill-WashCycle.wav", + "question": "does a light hissing noise start halfway through?", + "choice_a": "gas", + "choice_b": "chain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15627 + }, + { + "path": "Kali Temple Soundscape.wav", + "question": "Are these people swimming in water?", + "choice_a": "ball", + "choice_b": "seagulls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15628 + }, + { + "path": "squeaking wooden floor.wav", + "question": "How many times can a squeaking noise be heard?", + "choice_a": "three", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15629 + }, + { + "path": "Night Sounds in Orissa Village.wav", + "question": "How many people can be heard singing ?", + "choice_a": "rainy", + "choice_b": "bird", + "choice_c": "one", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15630 + }, + { + "path": "Drumming on a wine glass.wav", + "question": "Are there other sounds besides the tapping one?", + "choice_a": "six", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15631 + }, + { + "path": "bagpipe_on_street_BA.wav", + "question": "What type of instrument produces this sound?", + "choice_a": "clapping", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "canvas", + "answer_gt": "canvas", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15632 + }, + { + "path": "Little sreet behind a terrasse cafe.wav", + "question": "does the sound of the truck remain steady the whole time?", + "choice_a": "wrench", + "choice_b": "bell", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15633 + }, + { + "path": "Birds near pond.wav", + "question": "Is this taking place outside?", + "choice_a": "liquid", + "choice_b": "roof", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15634 + }, + { + "path": "Tiergarten birds early morning.wav", + "question": "are there many birds?", + "choice_a": "flushing", + "choice_b": "motor", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15635 + }, + { + "path": "Gibbons of Dusit.wav", + "question": "Does someone speak?", + "choice_a": "twenty", + "choice_b": "beeping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15636 + }, + { + "path": "Menziken Sawmill.wav", + "question": "What type of vehicle?", + "choice_a": "train", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15637 + }, + { + "path": "rhythm of the falling drops.wav", + "question": "What gender is the first person that talks?", + "choice_a": "men", + "choice_b": "flushing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "men", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15638 + }, + { + "path": "Hitting baseball w. wooden bat.wav", + "question": "Is there applause?", + "choice_a": "hall", + "choice_b": "clapping", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15639 + }, + { + "path": "fallingrice2.wav", + "question": "How often are objects thrown into the container?", + "choice_a": "slowly", + "choice_b": "car", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "slowly", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15640 + }, + { + "path": "Bizzare Atmosphere.wav", + "question": "Is banging occurring in the sound?", + "choice_a": "sparrow", + "choice_b": "seagulls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15641 + }, + { + "path": "clock_raw.wav", + "question": "how many beeps are there?", + "choice_a": "train", + "choice_b": "twentyfive", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "twentyfive", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15642 + }, + { + "path": "morning breeze and birds.wav", + "question": "Can a dog be heard barking?", + "choice_a": "stops", + "choice_b": "people", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15643 + }, + { + "path": "indoors house ambient room tone distant neighbours 1.wav", + "question": "Are the noises high-pitched?", + "choice_a": "dog", + "choice_b": "washing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15644 + }, + { + "path": "clock_raw.wav", + "question": "does the sound have a rhythm that is all over the place?", + "choice_a": "five", + "choice_b": "owl", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15645 + }, + { + "path": "cowshed.wav", + "question": "Are the cows mooing in a closed barn?", + "choice_a": "car", + "choice_b": "locust", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15646 + }, + { + "path": "20140210FallingIce.wav", + "question": "Is the person making popcorn?", + "choice_a": "one", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15647 + }, + { + "path": "bombolles.wav", + "question": "What is making the bubbling noise?", + "choice_a": "drill", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "liquid", + "answer_gt": "liquid", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15648 + }, + { + "path": "morning in the countryside.wav", + "question": "What is the female called, of the bird first heard?", + "choice_a": "thirty", + "choice_b": "yes", + "choice_c": "hen", + "choice_d": "no", + "answer_gt": "hen", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15649 + }, + { + "path": "CarInterior_RRx_WaitToTurn_FinalPark.wav", + "question": "Is this an engine?", + "choice_a": "fifteen", + "choice_b": "none", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15650 + }, + { + "path": "Night Sounds in Orissa Village.wav", + "question": "Is a man singing in the background ?", + "choice_a": "roof", + "choice_b": "heavy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15651 + }, + { + "path": "RG Large Old Dog Snoring.wav", + "question": "Is someone making a sound?", + "choice_a": "ten", + "choice_b": "five", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15652 + }, + { + "path": "Rain with thunder in a city.wav", + "question": "Is there a car alarm?", + "choice_a": "owl", + "choice_b": "drill", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15653 + }, + { + "path": "foley footsteps - raw.wav", + "question": "Is the recording device being moved at the start?", + "choice_a": "rainfall", + "choice_b": "canvas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15654 + }, + { + "path": "street_ambience_day.wav", + "question": "How many people can be heard clearly talking?", + "choice_a": "raining", + "choice_b": "beeping", + "choice_c": "three", + "choice_d": "no", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15655 + }, + { + "path": "Underwater_rumble_CsG.wav", + "question": "Does the rumbling ever stop?", + "choice_a": "sink", + "choice_b": "fast", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15656 + }, + { + "path": "20130418_stream.09.wav", + "question": "Does the sound get louder as time goes on?", + "choice_a": "slowly", + "choice_b": "flood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15657 + }, + { + "path": "Muddy_steps_bush_birds_singing.wav", + "question": "What is the person doing?", + "choice_a": "thunder", + "choice_b": "walking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15658 + }, + { + "path": "La Barca i La Tempesta.wav", + "question": "What animal is making noise?", + "choice_a": "mouse", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "mouse", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15659 + }, + { + "path": "interference from wireless mouse on am radio.wav", + "question": "What is buzzing?", + "choice_a": "low", + "choice_b": "static", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "static", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15660 + }, + { + "path": "Muddy_steps_bush_birds_singing.wav", + "question": "What is the person doing?", + "choice_a": "cars", + "choice_b": "snow", + "choice_c": "walking", + "choice_d": "no", + "answer_gt": "walking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15661 + }, + { + "path": "living room tone ambient distant noises neighbours.wav", + "question": "how many times is there a tap?", + "choice_a": "building", + "choice_b": "waves", + "choice_c": "zero", + "choice_d": "no", + "answer_gt": "zero", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15662 + }, + { + "path": "2012check_run.wav", + "question": "Are the formula one cars racing on the streets?", + "choice_a": "radio", + "choice_b": "drink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15663 + }, + { + "path": "Walking alongside the road.wav", + "question": "Do they open a door?", + "choice_a": "start", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15664 + }, + { + "path": "Atmosphere on road in London.wav", + "question": "Can a crowd of people be heard?", + "choice_a": "speaking", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15665 + }, + { + "path": "20090501.horse.neigh.wav", + "question": "Can the first animal that makes a noise be ridden?", + "choice_a": "bike", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15666 + }, + { + "path": "carpet_on_carpet.wav", + "question": "Are there multiple items being used?", + "choice_a": "coin", + "choice_b": "faucet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15667 + }, + { + "path": "forest_ambience_chepachet_spring_1.wav", + "question": "What are the birds doing?", + "choice_a": "sound", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "sound", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15668 + }, + { + "path": "020220_00.wav", + "question": "Does anyone respond to what the man is saying?", + "choice_a": "seagulls", + "choice_b": "bus", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15669 + }, + { + "path": "Fence Hit_City ambience night.wav", + "question": "Is the door made of wood?", + "choice_a": "party", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15670 + }, + { + "path": "Metal_On_Wood_Hits_Axe.wav", + "question": "does it sound like a wooden object is present?", + "choice_a": "tank", + "choice_b": "police", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15671 + }, + { + "path": "LondonTraffic.wav", + "question": "Is the trash truck moving down the road?", + "choice_a": "faucet", + "choice_b": "voice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15672 + }, + { + "path": "17-Year Cicada Mating Call.wav", + "question": "Is it an animal?", + "choice_a": "chicken", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15673 + }, + { + "path": "down stars running 3.wav", + "question": "What kind of noise is it?", + "choice_a": "water", + "choice_b": "band", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15674 + }, + { + "path": "Silencyo_RAIN_NIGHT_Village_CloseUp_Drips_Gutter_MSd.wav", + "question": "Is metal making noise?", + "choice_a": "never", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15675 + }, + { + "path": "windroar_constant_1m12s.wav", + "question": "How heavy is the rainfall?", + "choice_a": "barking", + "choice_b": "dog", + "choice_c": "heavy", + "choice_d": "no", + "answer_gt": "heavy", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15676 + }, + { + "path": "Cooking rice.wav", + "question": "Is a bus honking?", + "choice_a": "sheep", + "choice_b": "small", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15677 + }, + { + "path": "Birds in Pujipor.wav", + "question": "What is cawing?", + "choice_a": "running", + "choice_b": "normal", + "choice_c": "yes", + "choice_d": "crows", + "answer_gt": "crows", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15678 + }, + { + "path": "gargnano-sounds.wav", + "question": "Is there a dog?", + "choice_a": "glass", + "choice_b": "whispering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15679 + }, + { + "path": "Munich Cathedral.wav", + "question": "Is there music playing in the background?", + "choice_a": "voice", + "choice_b": "sink", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15680 + }, + { + "path": "gym machine.wav", + "question": "Is this apparatus operating speedily?", + "choice_a": "people", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15681 + }, + { + "path": "babbling brook 2.wav", + "question": "Is the liquid pouring?", + "choice_a": "twenty", + "choice_b": "hands", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15682 + }, + { + "path": "dragged-glass-object.wav", + "question": "Is someone using a hammer?", + "choice_a": "eight", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15683 + }, + { + "path": "small town.wav", + "question": "What is humming?", + "choice_a": "twenty", + "choice_b": "locust", + "choice_c": "driving", + "choice_d": "no", + "answer_gt": "driving", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15684 + }, + { + "path": "open and close pen.wav", + "question": "Is it a repetitive sound?", + "choice_a": "four", + "choice_b": "cleaning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15685 + }, + { + "path": "Hanoi street walking.wav", + "question": "are there any motors running?", + "choice_a": "distortion", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15686 + }, + { + "path": "STE-002-dishes_lisbon_restaurant.wav", + "question": "I there more than one person here?", + "choice_a": "once", + "choice_b": "rainfall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15687 + }, + { + "path": "AMB_EXT_PARK_SUMMER_DAY_LOOP.wav", + "question": "What other type of insect can make this sound?", + "choice_a": "locust", + "choice_b": "wrench", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "locust", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15688 + }, + { + "path": "Train Pass Koln.wav", + "question": "What is heard before the plane takes off?", + "choice_a": "loudspeaker", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "loudspeaker", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15689 + }, + { + "path": "070821_flsp_trail03.wav", + "question": "Does the airplane become more quiet over time?", + "choice_a": "ground", + "choice_b": "cow", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15690 + }, + { + "path": "birds_late_morning.wav", + "question": "What closure is opened at the end?", + "choice_a": "intercom", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "cabinet", + "answer_gt": "cabinet", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15691 + }, + { + "path": "eraser.wav", + "question": "Is there something metal clanking?", + "choice_a": "glass", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15692 + }, + { + "path": "crowd booing.wav", + "question": "Do the people seem upset?", + "choice_a": "windy", + "choice_b": "glass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15693 + }, + { + "path": "20160922_passing.lorry.marshes.wav", + "question": "Are there people talking to each other?", + "choice_a": "cricket", + "choice_b": "hall", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15694 + }, + { + "path": "footsteps on beach.wav", + "question": "Can people be heard in the cars driving by?", + "choice_a": "chicken", + "choice_b": "showering", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15695 + }, + { + "path": "Ambience - St Kilda Beach - waves lapping rocks, people nearby, seagulls.wav", + "question": "Is this near water?", + "choice_a": "siren", + "choice_b": "twentythree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15696 + }, + { + "path": "mercury topaz starting.wav", + "question": "Are the birds chirping?", + "choice_a": "glass", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15697 + }, + { + "path": "Japan_Tokyo_Shinjuku_Street_Promoter_Yelling_City.wav", + "question": "Is there distant music?", + "choice_a": "coin", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15698 + }, + { + "path": "Shower Running 02.wav", + "question": "Does the noise belong to the dripping water?", + "choice_a": "train", + "choice_b": "nothing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15699 + }, + { + "path": "Walking on dry grass.wav", + "question": "Is someone walking?", + "choice_a": "spinner", + "choice_b": "sand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15700 + }, + { + "path": "medical car horn EGYPT Alexandria.wav", + "question": "Is there more than one siren going off?", + "choice_a": "two", + "choice_b": "falling", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15701 + }, + { + "path": "Silencyo_CC_Field_Cicadas_French Children Playing_Wind_Dog_Distant Traffic.wav", + "question": "Were locust being heard?", + "choice_a": "hen", + "choice_b": "sand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15702 + }, + { + "path": "Howling Storm.wav", + "question": "how often does the whistle happen?", + "choice_a": "car", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "once", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15703 + }, + { + "path": "windy rain.wav", + "question": "Is it a wet day?", + "choice_a": "band", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15704 + }, + { + "path": "maryam sounds 5.wav", + "question": "is the jingling staying at the same pace the whole time?", + "choice_a": "outside", + "choice_b": "three", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15705 + }, + { + "path": "Inner City Bees.wav", + "question": "Are humans heard?", + "choice_a": "medium", + "choice_b": "cricket", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15706 + }, + { + "path": "tua-mirandela_train_arrival_march2007.wav", + "question": "What kind of animal is in the background?", + "choice_a": "cat", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15707 + }, + { + "path": "Rain and Storm.wav", + "question": "Is the rain hitting an umbrella?", + "choice_a": "siren", + "choice_b": "ground", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15708 + }, + { + "path": "door.of.bar.raining2.wav", + "question": "Arte men and women heard talking?", + "choice_a": "building", + "choice_b": "leaves", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15709 + }, + { + "path": "1400 am static.wav", + "question": "Do the wheels of the vehicle rotate on the asphalt?", + "choice_a": "umbrella", + "choice_b": "nothing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15710 + }, + { + "path": "WaterOnMetal.wav", + "question": "What is making the noise?", + "choice_a": "drink", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "machine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15711 + }, + { + "path": "Wind moaning through gap in door and house noises.wav", + "question": "How is the weather?", + "choice_a": "chirping", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rainy", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15712 + }, + { + "path": "Night Sounds in Orissa Village.wav", + "question": "Can loud thunder be heard ?", + "choice_a": "sand", + "choice_b": "canvas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15713 + }, + { + "path": "Stream # 2.wav", + "question": "Does the water flow without hindrances and obstructions?", + "choice_a": "once", + "choice_b": "honking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15714 + }, + { + "path": "Tree Bark Cracks.wav", + "question": "Is the person outside?", + "choice_a": "one", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15715 + }, + { + "path": "FOOTSTEPS_005.wav", + "question": "Is the person humming a tune?", + "choice_a": "road", + "choice_b": "driving", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15716 + }, + { + "path": "medical car horn EGYPT Alexandria.wav", + "question": "What type of siren is being used?", + "choice_a": "ambulance", + "choice_b": "once", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15717 + }, + { + "path": "branch and wind in wood 1.wav", + "question": "is a person making the sound?", + "choice_a": "coin", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15718 + }, + { + "path": "SPilling Water.wav", + "question": "Are there multiple types of noise?", + "choice_a": "fireworks", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15719 + }, + { + "path": "FP_Refrigerator_Door_Squeak.wav", + "question": "Is a door creaking?", + "choice_a": "sand", + "choice_b": "eating", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15720 + }, + { + "path": "forest.wav", + "question": "Does the music get louder?", + "choice_a": "clapping", + "choice_b": "driller", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15721 + }, + { + "path": "Scratching or tearing.wav", + "question": "Which instance of the scraping sound is the softest?", + "choice_a": "speaking", + "choice_b": "four", + "choice_c": "yes", + "choice_d": "paper", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15722 + }, + { + "path": "mercury topaz starting.wav", + "question": "Is the fan turned on?", + "choice_a": "traffic", + "choice_b": "party", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15723 + }, + { + "path": "Heavy Wind on Microphone.wav", + "question": "What speed is the wind blowing?", + "choice_a": "roof", + "choice_b": "fast", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "fast", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15724 + }, + { + "path": "Street Noise - Cars - Ball Bouncing indistinct voices.wav", + "question": "What animal was chirping?", + "choice_a": "eight", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15725 + }, + { + "path": "12 noon church-bell 140310_0121.wav", + "question": "What sound can be heard other than bird song?", + "choice_a": "cat", + "choice_b": "bell", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bell", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15726 + }, + { + "path": "Cardiff Bay fireworks.wav", + "question": "What is making the load bangs?", + "choice_a": "chopping", + "choice_b": "interested", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "interested", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15727 + }, + { + "path": "Flipping Coin Can.wav", + "question": "What kind of object is it?", + "choice_a": "insect", + "choice_b": "coin", + "choice_c": "paper", + "choice_d": "no", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15728 + }, + { + "path": "Electric Train Interior Atmos.wav", + "question": "Are there people talking in the background?", + "choice_a": "once", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15729 + }, + { + "path": "Underwater_rumble_CsG.wav", + "question": "Does the rumbling change tone at all?", + "choice_a": "gas", + "choice_b": "radio", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15730 + }, + { + "path": "walking indoors footsteps tap tap tapping foley.wav", + "question": "Where is this person walking?", + "choice_a": "four", + "choice_b": "constant", + "choice_c": "yes", + "choice_d": "hall", + "answer_gt": "hall", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15731 + }, + { + "path": "Street Noise - Cars - Ball Bouncing indistinct voices.wav", + "question": "What is the vehicle moving along?", + "choice_a": "water", + "choice_b": "clanging", + "choice_c": "bike", + "choice_d": "no", + "answer_gt": "bike", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15732 + }, + { + "path": "public.wav", + "question": "Are any birds making noise?", + "choice_a": "drilling", + "choice_b": "tap", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15733 + }, + { + "path": "Hamamatsu-traffic-light-1.wav", + "question": "the chiming usually happens when your car door is what?", + "choice_a": "outside", + "choice_b": "open", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "open", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15734 + }, + { + "path": "car acceleration sequence.wav", + "question": "how is the car started up in the beginning?", + "choice_a": "keys", + "choice_b": "falls", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "keys", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15735 + }, + { + "path": "Traffic Light.wav", + "question": "is it raining?", + "choice_a": "owl", + "choice_b": "crows", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15736 + }, + { + "path": "BobKessler-Spinning Tin Top.wav", + "question": "How many times was the thing rolled down?", + "choice_a": "building", + "choice_b": "zoo", + "choice_c": "four", + "choice_d": "no", + "answer_gt": "four", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15737 + }, + { + "path": "FOREST-SOUNDS.wav", + "question": "Are there at least two different types of bird sounds?", + "choice_a": "hall", + "choice_b": "screeching", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15738 + }, + { + "path": "food_prep_1_cw.wav", + "question": "Is someone clapping?", + "choice_a": "radio", + "choice_b": "twice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15739 + }, + { + "path": "Gulls at beach.wav", + "question": "Is there a large group of people?", + "choice_a": "engine", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15740 + }, + { + "path": "faucet3.wav", + "question": "how often does the sound continue?", + "choice_a": "dog", + "choice_b": "short", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "short", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15741 + }, + { + "path": "By ther blacksmith-002.wav", + "question": "Is something metallic being hit?", + "choice_a": "stops", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15742 + }, + { + "path": "bands_and_motorbike.wav", + "question": "How many times are the cymbals played?", + "choice_a": "constant", + "choice_b": "seagulls", + "choice_c": "yes", + "choice_d": "fifteen", + "answer_gt": "fifteen", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15743 + }, + { + "path": "Door Slam.wav", + "question": "Did a door close?", + "choice_a": "rainfall", + "choice_b": "machine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15744 + }, + { + "path": "Hallway Room Tone with shower in background.wav", + "question": "How many people can be heard?", + "choice_a": "three", + "choice_b": "none", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "none", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15745 + }, + { + "path": "Room-tone rain-drips 1m 161015_1013.wav", + "question": "Is the first sound longer than the middle and final sounds?", + "choice_a": "hall", + "choice_b": "cat", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15746 + }, + { + "path": "train.wav", + "question": "What type of vehicle makes this sound?", + "choice_a": "train", + "choice_b": "clanking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15747 + }, + { + "path": "1122thrum.wav", + "question": "How many times is there static?", + "choice_a": "two", + "choice_b": "vehicle", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15748 + }, + { + "path": "160917-eichelherr000.wav", + "question": "Is a bird making a bunch of noise?", + "choice_a": "water", + "choice_b": "chain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15749 + }, + { + "path": "Small growling dog.wav", + "question": "Is a dog growling ?", + "choice_a": "eating", + "choice_b": "announcement", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15750 + }, + { + "path": "Erny vs Deadman4.wav", + "question": "Are dogs barking at the people talking?", + "choice_a": "wind", + "choice_b": "road", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15751 + }, + { + "path": "Rain on Car Roof 2.wav", + "question": "Does it sound like damage is being done to property?", + "choice_a": "sweeping", + "choice_b": "twenty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15752 + }, + { + "path": "Alps village field-recording distance.wav", + "question": "Does it sound like cars are driving?", + "choice_a": "motorbike", + "choice_b": "loudspeaker", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15753 + }, + { + "path": "Alps village field-recording distance.wav", + "question": "Is a bird chirping in the background?", + "choice_a": "heavy", + "choice_b": "tearing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15754 + }, + { + "path": "tram_prague_2stops_veryfewpeople_AMB_INT.wav", + "question": "How many voices are heard?", + "choice_a": "bridge", + "choice_b": "bus", + "choice_c": "five", + "choice_d": "no", + "answer_gt": "five", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15755 + }, + { + "path": "steam train from 1912 locomotive.wav", + "question": "Can the rail tracks be heard clattering?", + "choice_a": "fireworks", + "choice_b": "buzzing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15756 + }, + { + "path": "Walking on dry grass.wav", + "question": "How many steps does the person take?", + "choice_a": "none", + "choice_b": "bell", + "choice_c": "yes", + "choice_d": "forty", + "answer_gt": "forty", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15757 + }, + { + "path": "cafecarusel_fan_hizz_EQ2.wav", + "question": "Does a cat meow repeatedly?", + "choice_a": "machine", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15758 + }, + { + "path": "mediterranean_sea_porticcio.wav", + "question": "Are there seagulls?", + "choice_a": "party", + "choice_b": "cabinet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15759 + }, + { + "path": "md3trk2.wav", + "question": "Is someone tearing up paper?", + "choice_a": "open", + "choice_b": "forty", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15760 + }, + { + "path": "bathroom drain 1.wav", + "question": "does the water sound remain steady in the same rhythm?", + "choice_a": "canvas", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15761 + }, + { + "path": "Liverpool St Station main hall.wav", + "question": "What is the person making an announcement into?", + "choice_a": "vehicle", + "choice_b": "drum", + "choice_c": "intercom", + "choice_d": "no", + "answer_gt": "intercom", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15762 + }, + { + "path": "Beach Wave Ambince .wav", + "question": "Where is the water coming from?", + "choice_a": "falls", + "choice_b": "eight", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "falls", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15763 + }, + { + "path": "weather_wind_strong_trees.wav", + "question": "Is it a calm day around?", + "choice_a": "six", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15764 + }, + { + "path": "Fryers Forest - Powerful Owl _Ninox Stenua_.wav", + "question": "What animal can be heard?", + "choice_a": "start", + "choice_b": "sparrow", + "choice_c": "yes", + "choice_d": "owl", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15765 + }, + { + "path": "Lluvia 1.wav", + "question": "If there is too much of this, what disaster can it cause?", + "choice_a": "normal", + "choice_b": "flood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "flood", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15766 + }, + { + "path": "FootstepsOnWoodFloor.wav", + "question": "Is a hammer being used?", + "choice_a": "thirty", + "choice_b": "rod", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15767 + }, + { + "path": "rhythm of the falling drops.wav", + "question": "Are there people speaking?", + "choice_a": "clanging", + "choice_b": "windy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15768 + }, + { + "path": "nightinggale2.wav", + "question": "Is this sound in the ocean?", + "choice_a": "road", + "choice_b": "wet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15769 + }, + { + "path": "Charleston Campus Summer.wav", + "question": "What are the people doing?", + "choice_a": "working", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "working", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15770 + }, + { + "path": "RBH_Household_shower 03.wav", + "question": "Is the person brushing their teeth?", + "choice_a": "clanking", + "choice_b": "no", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15771 + }, + { + "path": "Vomit, puking spilling water onto grass splat.wav", + "question": "What body part is the person using to pour the liquid out?", + "choice_a": "waves", + "choice_b": "hand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "hand", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15772 + }, + { + "path": "Rusty old boat.wav", + "question": "is nothing turned on?", + "choice_a": "one", + "choice_b": "dog", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15773 + }, + { + "path": "13. Crushing tin can.wav", + "question": "Are people walking in the background ?", + "choice_a": "fast", + "choice_b": "running", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15774 + }, + { + "path": "Rain and Storm.wav", + "question": "How many strikes of lightning can be heard?", + "choice_a": "one", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "one", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15775 + }, + { + "path": "Mug in sink.wav", + "question": "Is there a lighting strike?", + "choice_a": "wrench", + "choice_b": "zero", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15776 + }, + { + "path": "20080416.buzz.stereo.wav", + "question": "Are there birds chirping in the background ?", + "choice_a": "keys", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15777 + }, + { + "path": "Idling Truck at Loading Dock.wav", + "question": "Is the sound of the machine running far away?", + "choice_a": "sheep", + "choice_b": "chirp", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15778 + }, + { + "path": "steam train from 1912 locomotive.wav", + "question": "Can the rail tracks be heard clattering?", + "choice_a": "low", + "choice_b": "ball", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15779 + }, + { + "path": "Coffeehouse Ambience Burlington VT 0112xx.wav", + "question": "Are people around?", + "choice_a": "music", + "choice_b": "interested", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15780 + }, + { + "path": "20101228.teens.wav", + "question": "Are there children in the group?", + "choice_a": "zoo", + "choice_b": "fast", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15781 + }, + { + "path": "Car starting _open hood_.wav", + "question": "What is being started?", + "choice_a": "chopping", + "choice_b": "siren", + "choice_c": "car", + "choice_d": "no", + "answer_gt": "car", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15782 + }, + { + "path": "bird in the Hague at dawn 5.wav", + "question": "Is more than one bird chirping?", + "choice_a": "bird", + "choice_b": "closet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15783 + }, + { + "path": "Morsecode - SOS MAYDAY - 988 Hz Tone.wav", + "question": "Is a machine making this noise?", + "choice_a": "siren", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15784 + }, + { + "path": "LA Rain.wav", + "question": "How many footsteps can be heard in the clip?", + "choice_a": "keys", + "choice_b": "ground", + "choice_c": "zero", + "choice_d": "no", + "answer_gt": "zero", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15785 + }, + { + "path": "bridge demolition pounding.wav", + "question": "Is this in a warehouse?", + "choice_a": "eating", + "choice_b": "umbrella", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15786 + }, + { + "path": "Gravel_Sand Walking 1.wav", + "question": "Is a person walking on soft ground?", + "choice_a": "iron", + "choice_b": "rainy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15787 + }, + { + "path": "01 residential elevator with door open and close.wav", + "question": "Is the light roar?", + "choice_a": "drill", + "choice_b": "spinner", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15788 + }, + { + "path": "Waterfalls_00216.wav", + "question": "Is the spraying sound loud?", + "choice_a": "bridge", + "choice_b": "announcement", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15789 + }, + { + "path": "gym machine.wav", + "question": "Is the sound a voice?", + "choice_a": "none", + "choice_b": "start", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15790 + }, + { + "path": "Kaffemaschine_1.wav", + "question": "What is been used?", + "choice_a": "engine", + "choice_b": "door", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "engine", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15791 + }, + { + "path": "Davis.wav", + "question": "If one were caught outside, how would they end up being?", + "choice_a": "drink", + "choice_b": "start", + "choice_c": "wet", + "choice_d": "no", + "answer_gt": "wet", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15792 + }, + { + "path": "a gentle breeze, wind 4.wav", + "question": "Is the sound produced by inanimate objects?", + "choice_a": "rain", + "choice_b": "voice", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15793 + }, + { + "path": "International Harvester Scout II.wav", + "question": "Is something being smacked?", + "choice_a": "window", + "choice_b": "metal", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15794 + }, + { + "path": "Library Ambience_large space.wav", + "question": "What sort of noise coming from a human can briefly be heard in the background?", + "choice_a": "waves", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "talking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15795 + }, + { + "path": "Squeaky Wood _Compilation_.wav", + "question": "Is someone yelling?", + "choice_a": "seagulls", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15796 + }, + { + "path": "Bubbles water.wav", + "question": "Is the liquid frozen?", + "choice_a": "snow", + "choice_b": "distortion", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15797 + }, + { + "path": "small dog leaves.wav", + "question": "What is being crumpled?", + "choice_a": "showering", + "choice_b": "paper", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15798 + }, + { + "path": "110724_inriversidemus1.wav", + "question": "Is the person speaking outside?", + "choice_a": "paper", + "choice_b": "two", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15799 + }, + { + "path": "tornado day 1.wav", + "question": "What is the weather like?", + "choice_a": "raining", + "choice_b": "open", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "raining", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15800 + }, + { + "path": "chrrrckkkcc.wav", + "question": "Is this happening inside?", + "choice_a": "ten", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15801 + }, + { + "path": "Footsteps, Muddy, E.wav", + "question": "Is the person walking?", + "choice_a": "drink", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15802 + }, + { + "path": "AC Unit.wav", + "question": "What are they working on?", + "choice_a": "driller", + "choice_b": "engine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "driller", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15803 + }, + { + "path": "Dinosaur Footsteps-01.wav", + "question": "How many thumps take place?", + "choice_a": "static", + "choice_b": "four", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "four", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15804 + }, + { + "path": "Erny vs Deadman4.wav", + "question": "are there more than two person in the arguing?", + "choice_a": "insect", + "choice_b": "ten", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15805 + }, + { + "path": "CalmWaves SandBeach 03 EQ 130430_03.wav", + "question": "What is making the noise?", + "choice_a": "waves", + "choice_b": "locust", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "waves", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15806 + }, + { + "path": "20130418_stream.09.wav", + "question": "Does the sound get louder as time goes on?", + "choice_a": "yes", + "choice_b": "tank", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15807 + }, + { + "path": "earth_movement.wav", + "question": "Does the motor stay idle the whole time?", + "choice_a": "vehicle", + "choice_b": "sound", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15808 + }, + { + "path": "Cityscape construction site 2 100304.wav", + "question": "Is someone playing the trumpet ?", + "choice_a": "eight", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15809 + }, + { + "path": "Face slap CsG.wav", + "question": "What item is being hit against the object?", + "choice_a": "mixer", + "choice_b": "bear", + "choice_c": "hand", + "choice_d": "no", + "answer_gt": "hand", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15810 + }, + { + "path": "Streatham Railway Station and on a Train.wav", + "question": "What form of transportation is this?", + "choice_a": "bus", + "choice_b": "honking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15811 + }, + { + "path": "walkingondirtpath.wav", + "question": "Is the man walking on pebbles?", + "choice_a": "ocean", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15812 + }, + { + "path": "INT Factory budimka, pozega.wav", + "question": "What type of work would make these sounds?", + "choice_a": "fast", + "choice_b": "walking", + "choice_c": "washing", + "choice_d": "no", + "answer_gt": "washing", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15813 + }, + { + "path": "drunk_teenagers_1.wav", + "question": "who is talking besides men?", + "choice_a": "low", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "women", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15814 + }, + { + "path": "20101228.teens.wav", + "question": "Is somebody giving a speech?", + "choice_a": "drill", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15815 + }, + { + "path": "Car starting _open hood_.wav", + "question": "Is this the sound of a car?", + "choice_a": "steps", + "choice_b": "keys", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15816 + }, + { + "path": "ROLDAN_ANDRES_GEO_FONIA_AGUA_DUCHA.wav", + "question": "Is anyone talking?", + "choice_a": "zoo", + "choice_b": "birds", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15817 + }, + { + "path": "h907 boules pologna clap f.wav", + "question": "Why are the people applauding?", + "choice_a": "announcement", + "choice_b": "music", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "announcement", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15818 + }, + { + "path": "Rain drops on marquee.wav", + "question": "is there thunder?", + "choice_a": "one", + "choice_b": "glass", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15819 + }, + { + "path": "Ship Fender.wav", + "question": "Does the engine keep cutting out?", + "choice_a": "clapping", + "choice_b": "six", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15820 + }, + { + "path": "Horse_Hooves_Hard_Floor_Interior.wav", + "question": "how many sets of clomping noises are there?", + "choice_a": "heavy", + "choice_b": "crows", + "choice_c": "yes", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15821 + }, + { + "path": "storm is coming 15-11-2012.wav", + "question": "Is it raining?", + "choice_a": "sparrow", + "choice_b": "running", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15822 + }, + { + "path": "Watervogels en riet Lichterveldestraat _HiPass_.wav", + "question": "Are there any voices in the background?", + "choice_a": "beeping", + "choice_b": "fifteen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15823 + }, + { + "path": "md1trk33-34.wav", + "question": "Is there a quick fix to remedy this sound?", + "choice_a": "loudspeaker", + "choice_b": "women", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15824 + }, + { + "path": "2013-03-28 rain in the rainforest.wav", + "question": "Is the sound continuous?", + "choice_a": "swimming", + "choice_b": "insect", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15825 + }, + { + "path": "steam engine at museum.wav", + "question": "Are there tree leaves being raked up?", + "choice_a": "roof", + "choice_b": "eight", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15826 + }, + { + "path": "STE-008.wav", + "question": "What sound is the train making?", + "choice_a": "intercom", + "choice_b": "honking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "honking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15827 + }, + { + "path": "Arch Leaf.wav", + "question": "Are the footsteps on a waxed floor?", + "choice_a": "lightning", + "choice_b": "flushing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15828 + }, + { + "path": "thunder2.wav", + "question": "How is the weather?", + "choice_a": "fast", + "choice_b": "rain", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15829 + }, + { + "path": "20110805_forest.crows.07.wav", + "question": "Is an animal making noise?", + "choice_a": "buzzsaw", + "choice_b": "outside", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15830 + }, + { + "path": "International Harvester Scout II.wav", + "question": "Is a car making noise?", + "choice_a": "building", + "choice_b": "announcement", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15831 + }, + { + "path": "Scratching or tearing.wav", + "question": "Is there an elk?", + "choice_a": "nothing", + "choice_b": "one", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15832 + }, + { + "path": "Kung Fu Clothes Hits and Clothing Sounds.wav", + "question": "What is hit?", + "choice_a": "outdoors", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "ball", + "answer_gt": "ball", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15833 + }, + { + "path": "abandoned-ballroom-big-metal.wav", + "question": "Is it raining?", + "choice_a": "running", + "choice_b": "swimming", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15834 + }, + { + "path": "Theater Chatter.wav", + "question": "What are the people doing?", + "choice_a": "bridge", + "choice_b": "sink", + "choice_c": "talking", + "choice_d": "no", + "answer_gt": "talking", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15835 + }, + { + "path": "SpringPeepersMarch2012.wav", + "question": "Is this animal noises?", + "choice_a": "two", + "choice_b": "steps", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15836 + }, + { + "path": "FOOTSTEPS_005.wav", + "question": "Is the person humming a tune?", + "choice_a": "whispering", + "choice_b": "cutter", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15837 + }, + { + "path": "foley footsteps - raw.wav", + "question": "Are there multiple people present?", + "choice_a": "drum", + "choice_b": "building", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15838 + }, + { + "path": "20081130_walking_in_snow_with_snowshoes.wav", + "question": "Is the person in a cold environment?", + "choice_a": "chirp", + "choice_b": "bird", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15839 + }, + { + "path": "Ronda - Fountain near the Town Hall _general_ - Fuente cerca del Ayuntamiento _general_.wav", + "question": "What object is sometimes carried by a person to stay dry when this is happening?", + "choice_a": "loudspeaker", + "choice_b": "umbrella", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "umbrella", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15840 + }, + { + "path": "Sink_Running2.wav", + "question": "Are there footsteps?", + "choice_a": "tank", + "choice_b": "stops", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15841 + }, + { + "path": "HammerDrill.wav", + "question": "How many times does the wooden figure croak?", + "choice_a": "six", + "choice_b": "chirp", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "six", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15842 + }, + { + "path": "Cooking on Gas.wav", + "question": "does the sound come from nature?", + "choice_a": "singing", + "choice_b": "windy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15843 + }, + { + "path": "audience final applause 01.wav", + "question": "does the clapping ever cease?", + "choice_a": "rod", + "choice_b": "talking", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15844 + }, + { + "path": "Elysian Park - Picnic Area 2.wav", + "question": "Is the road very busy?", + "choice_a": "forty", + "choice_b": "heavy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15845 + }, + { + "path": "NY subway.wav", + "question": "Are there any females?", + "choice_a": "once", + "choice_b": "nine", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15846 + }, + { + "path": "Crunching Grass.wav", + "question": "how many footsteps are there?", + "choice_a": "sweeping", + "choice_b": "twelve", + "choice_c": "yes", + "choice_d": "twentythree", + "answer_gt": "twentythree", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15847 + }, + { + "path": "House_kettle boil_whistle.wav", + "question": "Does the pitch of the whistle get higher?", + "choice_a": "yes", + "choice_b": "slowly", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15848 + }, + { + "path": "CR FunnyMachine.wav", + "question": "The machine is making a high-pitched noise referred to as what?", + "choice_a": "drill", + "choice_b": "liquid", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "drill", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15849 + }, + { + "path": "Thunder Storm Daytona 3.wav", + "question": "What was moved?", + "choice_a": "birds", + "choice_b": "storm", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "storm", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15850 + }, + { + "path": "Creaky wooden steps, down and up.wav", + "question": "Do these sounds involve human movement?", + "choice_a": "sheep", + "choice_b": "gas", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15851 + }, + { + "path": "greece_naxos_cicadas_4.wav", + "question": "what animal is making the sound?", + "choice_a": "snow", + "choice_b": "drill", + "choice_c": "yes", + "choice_d": "cricket", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15852 + }, + { + "path": "CityPark Evening Moerputten NL 130510_01.wav", + "question": "Is there more than one bird chirping?", + "choice_a": "stops", + "choice_b": "yes", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15853 + }, + { + "path": "Inner City Bees.wav", + "question": "Does the buzzing intensify?", + "choice_a": "working", + "choice_b": "singing", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15854 + }, + { + "path": "DR0000_0020.wav", + "question": "is there a large storm?", + "choice_a": "birds", + "choice_b": "saw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15855 + }, + { + "path": "pencil on paper.wav", + "question": "People commonly do this action upon a sheet of what?", + "choice_a": "driller", + "choice_b": "paper", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "paper", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15856 + }, + { + "path": "Distorted AM Radio noise.wav", + "question": "Is a person singing?", + "choice_a": "stops", + "choice_b": "working", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15857 + }, + { + "path": "wheaten field.wav", + "question": "What is hissing?", + "choice_a": "forty", + "choice_b": "lightning", + "choice_c": "yes", + "choice_d": "wind", + "answer_gt": "wind", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15858 + }, + { + "path": "Box of sweets.wav", + "question": "Is the person eating?", + "choice_a": "iron", + "choice_b": "never", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15859 + }, + { + "path": "Beer Pong Sounds ball table and cups.wav", + "question": "Does the object fall at the beginning?", + "choice_a": "electronic", + "choice_b": "wood", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15860 + }, + { + "path": "DoorSqueak.wav", + "question": "Is something creaking?", + "choice_a": "bike", + "choice_b": "sand", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15861 + }, + { + "path": "pr#6F9A9E.wav", + "question": "Is something dropping?", + "choice_a": "medium", + "choice_b": "wet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15862 + }, + { + "path": "Airplane Overhead.wav", + "question": "Is this a form of transportation?", + "choice_a": "ocean", + "choice_b": "brush", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15863 + }, + { + "path": "Kiddie Train.wav", + "question": "Is there a horn making sound?", + "choice_a": "wind", + "choice_b": "wet", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15864 + }, + { + "path": "Night Ambient.wav", + "question": "Is there thunder?", + "choice_a": "outside", + "choice_b": "buzzsaw", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "yes", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15865 + }, + { + "path": "Cooking on Gas.wav", + "question": "when does the sound cease?", + "choice_a": "never", + "choice_b": "hen", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "never", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15866 + }, + { + "path": "2012check_run.wav", + "question": "How many vehicles are there?", + "choice_a": "two", + "choice_b": "twentythree", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "two", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15867 + }, + { + "path": "Footsteps On Squeaky Wood Floor.wav", + "question": "Is the person walking barefoot?", + "choice_a": "music", + "choice_b": "windy", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15868 + }, + { + "path": "Single cricket chirping during a summer evening in the city _with traffic noise_.wav", + "question": "Is someone pouring out a liquid?", + "choice_a": "wrench", + "choice_b": "water", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15869 + }, + { + "path": "h907 boules pologna clap f.wav", + "question": "Is this outside?", + "choice_a": "working", + "choice_b": "lock", + "choice_c": "yes", + "choice_d": "no", + "answer_gt": "no", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15870 + }, + { + "path": "bar crowd.wav", + "question": "Besides loud speech, what other human sound can be heard?", + "choice_a": "roof", + "choice_b": "motor", + "choice_c": "yes", + "choice_d": "person", + "answer_gt": "person", + "task_name": "Sound_AQA", + "dataset_name": "clothoaqa", + "uniq_id": 15871 + }, + { + "path": "6236.flac", + "question": "What animal makes a sound in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "cat", + "choice_d": "bird", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15872 + }, + { + "path": "2879.flac", + "question": "What is the main sound source of the video?", + "choice_a": "automobile", + "choice_b": "aircraft", + "choice_c": "metro", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15873 + }, + { + "path": "3511.flac", + "question": "What's the weather in the video?", + "choice_a": "a furious storm", + "choice_b": "snow", + "choice_c": "the wind is mild and the sun is bright", + "choice_d": "sand storm", + "answer_gt": "a furious storm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15874 + }, + { + "path": "37868.flac", + "question": "Where did the video take place?", + "choice_a": "seabed", + "choice_b": "playground", + "choice_c": "zoo", + "choice_d": "beach", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15875 + }, + { + "path": "37547.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15876 + }, + { + "path": "37202.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Fire engine siren", + "choice_b": "Ambulance siren", + "choice_c": "Car horn", + "choice_d": "honk", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15877 + }, + { + "path": "45166.flac", + "question": "What happened in the video?", + "choice_a": "Birds break their shells", + "choice_b": "to brood", + "choice_c": "Birds caught", + "choice_d": "Outing", + "answer_gt": "Birds break their shells", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15878 + }, + { + "path": "28305.flac", + "question": "Where does the video take place?", + "choice_a": "playground", + "choice_b": "kindergarten", + "choice_c": "Playground", + "choice_d": "natatorium", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15879 + }, + { + "path": "60257.flac", + "question": "What is shown in the video?", + "choice_a": "Game console keyboard", + "choice_b": "Typing keyboard", + "choice_c": "Radio translation", + "choice_d": "Coding machine", + "answer_gt": "Typing keyboard", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15880 + }, + { + "path": "60780.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "hare", + "choice_c": "kitten", + "choice_d": "timber wolf", + "answer_gt": "kitten", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15881 + }, + { + "path": "59151.flac", + "question": "What is shown in the video?", + "choice_a": "Motorboat acceleration", + "choice_b": "Sailing", + "choice_c": "Dragon Boat Race", + "choice_d": "Cargo ship sailing", + "answer_gt": "Sailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15882 + }, + { + "path": "39895.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Remote control train", + "choice_b": "remote controlled aircraft", + "choice_c": "Remote control car", + "choice_d": "Remote control car", + "answer_gt": "remote controlled aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15883 + }, + { + "path": "31796.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "The swallow calls", + "choice_b": "The magpie calls", + "choice_c": "The pigeon calls", + "choice_d": "Crows crow", + "answer_gt": "The swallow calls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15884 + }, + { + "path": "22896.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "duck", + "choice_c": "chicken", + "choice_d": "dove", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15885 + }, + { + "path": "7232.flac", + "question": "Where does the video take place?", + "choice_a": "In the park", + "choice_b": "Toilet", + "choice_c": "On the street", + "choice_d": "courtyard", + "answer_gt": "courtyard", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15886 + }, + { + "path": "33585.flac", + "question": "Where did the video take place?", + "choice_a": "seabed", + "choice_b": "playground", + "choice_c": "court", + "choice_d": "Ski Field", + "answer_gt": "Ski Field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15887 + }, + { + "path": "10720.flac", + "question": "What are the people doing in the video?", + "choice_a": "Gun fire", + "choice_b": "performing oral skills", + "choice_c": "Knock", + "choice_d": "Lighter to light", + "answer_gt": "Gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15888 + }, + { + "path": "5529.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "motorboat", + "choice_c": "kite", + "choice_d": "aircraft", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15889 + }, + { + "path": "34842.flac", + "question": "What's spinning in the video?", + "choice_a": "propeller", + "choice_b": "Fan", + "choice_c": "air conditioner", + "choice_d": "windmill", + "answer_gt": "windmill", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15890 + }, + { + "path": "26480.flac", + "question": "What are the people in the video doing?", + "choice_a": "rope skipping", + "choice_b": "beat a gong", + "choice_c": "practise qigong", + "choice_d": "Nailing", + "answer_gt": "Nailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15891 + }, + { + "path": "51621.flac", + "question": "Why did the car in the video stop?", + "choice_a": "Observe the lion", + "choice_b": "Take a picture of the lion", + "choice_c": "The lion is in the way", + "choice_d": "Scared by the lion", + "answer_gt": "The lion is in the way", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15892 + }, + { + "path": "27138.flac", + "question": "What are the people doing in the video?", + "choice_a": "run", + "choice_b": "skiing", + "choice_c": "Mountaineering", + "choice_d": "skating", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15893 + }, + { + "path": "27653.flac", + "question": "What happened in the video?", + "choice_a": "Reversing", + "choice_b": "light firecrackers", + "choice_c": "Repair the tire", + "choice_d": "Cars crush things", + "answer_gt": "Cars crush things", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15894 + }, + { + "path": "58012.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Machining parts sound", + "choice_b": "Grinding noise of lathe", + "choice_c": "Lathe rotation sound", + "choice_d": "Making ceramics", + "answer_gt": "Lathe rotation sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15895 + }, + { + "path": "51961.flac", + "question": "What color is the woman's hair in the video?", + "choice_a": "yellow", + "choice_b": "red", + "choice_c": "Grape violet", + "choice_d": "Light brown", + "answer_gt": "Light brown", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15896 + }, + { + "path": "1498.flac", + "question": "What are the people doing in the video?", + "choice_a": "Shooting", + "choice_b": "Playing games", + "choice_c": "speak", + "choice_d": "anchor", + "answer_gt": "Playing games", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15897 + }, + { + "path": "37272.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "sparrow", + "choice_c": "dove", + "choice_d": "chick", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15898 + }, + { + "path": "10830.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "fox", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15899 + }, + { + "path": "33712.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of wind", + "choice_b": "sound of rain", + "choice_c": "aircraft", + "choice_d": "thunder", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15900 + }, + { + "path": "16682.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "cock", + "choice_b": "dog", + "choice_c": "dove", + "choice_d": "duck", + "answer_gt": "cock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15901 + }, + { + "path": "6248.flac", + "question": "What animal is in the video?", + "choice_a": "honeybee", + "choice_b": "owl", + "choice_c": "swallow", + "choice_d": "parrot", + "answer_gt": "parrot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15902 + }, + { + "path": "4451.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "rabbit", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15903 + }, + { + "path": "40002.flac", + "question": "Where does the video take place?", + "choice_a": "seabed", + "choice_b": "land", + "choice_c": "in the air", + "choice_d": "In the sea", + "answer_gt": "in the air", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15904 + }, + { + "path": "13454.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "motorcycle", + "choice_d": "metro", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15905 + }, + { + "path": "54592.flac", + "question": "Where did the video take place?", + "choice_a": "farm", + "choice_b": "Golf Course", + "choice_c": "tennis court", + "choice_d": "Bowling alley", + "answer_gt": "Bowling alley", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15906 + }, + { + "path": "26196.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "sheep", + "choice_d": "dog", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15907 + }, + { + "path": "53079.flac", + "question": "Where did the video take place?", + "choice_a": "park", + "choice_b": "Paddock", + "choice_c": "farm", + "choice_d": "Farmyard", + "answer_gt": "farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15908 + }, + { + "path": "3030.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "fox", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15909 + }, + { + "path": "3593.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "ambulance", + "choice_c": "police car", + "choice_d": "Motorcycles", + "answer_gt": "police car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15910 + }, + { + "path": "36346.flac", + "question": "What are the people doing in the video?", + "choice_a": "Driving a ship", + "choice_b": "Driving a motorcycle", + "choice_c": "Driving a tractor", + "choice_d": "Driving a bus", + "answer_gt": "Driving a bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15911 + }, + { + "path": "5879.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "waterfall", + "choice_b": "water", + "choice_c": "sound of wind", + "choice_d": "Machine sound", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15912 + }, + { + "path": "61633.flac", + "question": "What are the people doing in the video?", + "choice_a": "Play basketball", + "choice_b": "Bowling", + "choice_c": "Play baseball", + "choice_d": "Practice shot put", + "answer_gt": "Bowling", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15913 + }, + { + "path": "34661.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Eat less", + "choice_b": "Eat less", + "choice_c": "motorcycle", + "choice_d": "Eat less", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15914 + }, + { + "path": "27532.flac", + "question": "What are the people in the video doing?", + "choice_a": "Driving a motorcycle", + "choice_b": "Driving a racing car", + "choice_c": "Driving a train", + "choice_d": "Driving a go kart", + "answer_gt": "Driving a racing car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15915 + }, + { + "path": "25153.flac", + "question": "What are the people doing in the video?", + "choice_a": "Nailing", + "choice_b": "Cut wood", + "choice_c": "rope skipping", + "choice_d": "knock at the door", + "answer_gt": "Nailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15916 + }, + { + "path": "18446.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "metro", + "choice_c": "train", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15917 + }, + { + "path": "58135.flac", + "question": "What happened in the video?", + "choice_a": "rowing", + "choice_b": "to play", + "choice_c": "Dragon-boat Racing", + "choice_d": "Playing in a motorboat", + "answer_gt": "Dragon-boat Racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15918 + }, + { + "path": "3696.flac", + "question": "What happened in the video?", + "choice_a": "pump", + "choice_b": "Volcanic explosion", + "choice_c": "firecrackers", + "choice_d": "cars crush things", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15919 + }, + { + "path": "8806.flac", + "question": "What animal is in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "donkey", + "choice_d": "bird", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15920 + }, + { + "path": "34413.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "hail", + "choice_c": "sound of rain", + "choice_d": "thunder", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15921 + }, + { + "path": "39850.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "metro", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15922 + }, + { + "path": "7056.flac", + "question": "What animal is in the video?", + "choice_a": "lion", + "choice_b": "dog", + "choice_c": "wolf", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15923 + }, + { + "path": "55935.flac", + "question": "What happened in the video?", + "choice_a": "fire", + "choice_b": "Burn a fire", + "choice_c": "The house is on fire", + "choice_d": "warm oneself", + "answer_gt": "Burn a fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15924 + }, + { + "path": "38321.flac", + "question": "What animals appear in the video?", + "choice_a": "dove", + "choice_b": "crow", + "choice_c": "parrot", + "choice_d": "duck", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15925 + }, + { + "path": "2123.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "metro", + "choice_b": "train", + "choice_c": "highway", + "choice_d": "bus", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15926 + }, + { + "path": "31550.flac", + "question": "What animal appears in the video?", + "choice_a": "Canary", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "dog", + "answer_gt": "Canary", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15927 + }, + { + "path": "40527.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorcycle", + "choice_b": "train", + "choice_c": "the sound of driving a snowmobile", + "choice_d": "lawn mower", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15928 + }, + { + "path": "24301.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "fox", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15929 + }, + { + "path": "42764.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "shot", + "choice_b": "rope skipping", + "choice_c": "firecrackers", + "choice_d": "Announcer", + "answer_gt": "shot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15930 + }, + { + "path": "40062.flac", + "question": "Where does the video take place?", + "choice_a": "of one's own unit", + "choice_b": "racetrack", + "choice_c": "surface of the water", + "choice_d": "seabed", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15931 + }, + { + "path": "35234.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "owl", + "choice_b": "eagle", + "choice_c": "wild goose", + "choice_d": "tiger", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15932 + }, + { + "path": "22472.flac", + "question": "What animal is in the video?", + "choice_a": "crocodile", + "choice_b": "snake", + "choice_c": "chameleon", + "choice_d": "chopped food", + "answer_gt": "crocodile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15933 + }, + { + "path": "7331.flac", + "question": "Where did the video take place?", + "choice_a": "field", + "choice_b": "seabed", + "choice_c": "ski field", + "choice_d": "at sea", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15934 + }, + { + "path": "26573.flac", + "question": "Where did the video take place?", + "choice_a": "highway", + "choice_b": "playground", + "choice_c": "in the air", + "choice_d": "aquatic", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15935 + }, + { + "path": "24872.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "helicopter", + "choice_b": "automobile", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15936 + }, + { + "path": "3515.flac", + "question": "What are the people doing in this video?", + "choice_a": "machine", + "choice_b": "rowing", + "choice_c": "skiing", + "choice_d": "Rope skipping", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15937 + }, + { + "path": "43003.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Car horn", + "choice_b": "Machine sound", + "choice_c": "Airplane noise", + "choice_d": "Bus horn", + "answer_gt": "Car horn", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15938 + }, + { + "path": "28329.flac", + "question": "What animal appears in the video?", + "choice_a": "fox", + "choice_b": "dog", + "choice_c": "wolf", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15939 + }, + { + "path": "1297.flac", + "question": "What's running in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15940 + }, + { + "path": "49035.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "truck", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15941 + }, + { + "path": "44098.flac", + "question": "Where did the video take place?", + "choice_a": "farm", + "choice_b": "toilet", + "choice_c": "pasture", + "choice_d": "zoo", + "answer_gt": "farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15942 + }, + { + "path": "5904.flac", + "question": "What are the people doing in the video?", + "choice_a": "sow", + "choice_b": "Weeding", + "choice_c": "motorboat", + "choice_d": "plough fields", + "answer_gt": "Weeding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15943 + }, + { + "path": "51184.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "120 ambulance", + "choice_b": "Police car siren", + "choice_c": "whistling", + "choice_d": "Go kart horn", + "answer_gt": "Police car siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15944 + }, + { + "path": "41089.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "train", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15945 + }, + { + "path": "34438.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "sound of wind", + "choice_c": "aircraft", + "choice_d": "Bus", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15946 + }, + { + "path": "32654.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Civil defense alarm", + "choice_b": "sound of wind", + "choice_c": "Train horn", + "choice_d": "motorboat", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15947 + }, + { + "path": "48341.flac", + "question": "Where did the video take place?", + "choice_a": "in the mountains", + "choice_b": "aquatic", + "choice_c": "shore", + "choice_d": "Bathing ducks", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15948 + }, + { + "path": "60389.flac", + "question": "What are the people doing in the video?", + "choice_a": "Folding paper", + "choice_b": "Check out the list", + "choice_c": "Sorting out receipts", + "choice_d": "Finishing paper", + "answer_gt": "Finishing paper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15949 + }, + { + "path": "51080.flac", + "question": "What's the view like in the video?", + "choice_a": "a furious storm", + "choice_b": "the wind is mild and the sun is bright", + "choice_c": "The smoke is floating", + "choice_d": "charming", + "answer_gt": "the wind is mild and the sun is bright", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15950 + }, + { + "path": "15125.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "ambulance", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15951 + }, + { + "path": "21964.flac", + "question": "Where did the video take place?", + "choice_a": "field", + "choice_b": "ski field", + "choice_c": "zoo", + "choice_d": "Playground", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15952 + }, + { + "path": "14813.flac", + "question": "What is shown in the video?", + "choice_a": "The ball", + "choice_b": "Football", + "choice_c": "earth", + "choice_d": "Space", + "answer_gt": "earth", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15953 + }, + { + "path": "50480.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Car horn", + "choice_b": "Police car siren", + "choice_c": "voice", + "choice_d": "Bus horn", + "answer_gt": "Police car siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15954 + }, + { + "path": "12187.flac", + "question": "Where does the video take place?", + "choice_a": "On the mountain", + "choice_b": "aquatic", + "choice_c": "field", + "choice_d": "in the air", + "answer_gt": "On the mountain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15955 + }, + { + "path": "21103.flac", + "question": "What are the people in the video doing?", + "choice_a": "Stir fry vegetables", + "choice_b": "cut up vegetables", + "choice_c": "eat fruit", + "choice_d": "Cutting food", + "answer_gt": "Cutting food", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15956 + }, + { + "path": "56830.flac", + "question": "What happened in the video?", + "choice_a": "Scanner scanning", + "choice_b": "Printer printing", + "choice_c": "Lettering machine", + "choice_d": "Press printing", + "answer_gt": "Printer printing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15957 + }, + { + "path": "49425.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "aircraft", + "choice_c": "Tractor", + "choice_d": "Bulldozer", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15958 + }, + { + "path": "19319.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "car engine noise", + "choice_c": "aircraft", + "choice_d": "vacuum cleaner to clean the floor", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15959 + }, + { + "path": "47913.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "aircraft", + "choice_b": "Tractor", + "choice_c": "train", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15960 + }, + { + "path": "52541.flac", + "question": "What are the people doing in the video?", + "choice_a": "practice oral skills", + "choice_b": "play table tennis", + "choice_c": "Playing ice hockey", + "choice_d": "play volleyball", + "answer_gt": "play volleyball", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15961 + }, + { + "path": "6343.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "parrot", + "choice_b": "peacock", + "choice_c": "dove", + "choice_d": "cock", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15962 + }, + { + "path": "39137.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "park", + "choice_c": "seabed", + "choice_d": "aquatic", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15963 + }, + { + "path": "30604.flac", + "question": "What are the people doing in the video?", + "choice_a": "run", + "choice_b": "dance", + "choice_c": "Rope skipping", + "choice_d": "long jump", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15964 + }, + { + "path": "17412.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "bird", + "choice_c": "Hyenas", + "choice_d": "fox", + "answer_gt": "Hyenas", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15965 + }, + { + "path": "30292.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "duck", + "choice_c": "cat", + "choice_d": "chicken", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15966 + }, + { + "path": "11121.flac", + "question": "What animal is in the video?", + "choice_a": "owl", + "choice_b": "bird", + "choice_c": "chicken", + "choice_d": "duck", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15967 + }, + { + "path": "32121.flac", + "question": "What animals appear in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "cattle", + "choice_d": "donkey", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15968 + }, + { + "path": "40199.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorcycle", + "choice_b": "Car", + "choice_c": "train", + "choice_d": "the sound of driving a snowmobile", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15969 + }, + { + "path": "23451.flac", + "question": "What animal is in the video?", + "choice_a": "planing", + "choice_b": "mosquito", + "choice_c": "fireworks and firecrackers", + "choice_d": "honeybee", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15970 + }, + { + "path": "17705.flac", + "question": "What animal appears in the video?", + "choice_a": "duck", + "choice_b": "swallow", + "choice_c": "rabbit", + "choice_d": "goose", + "answer_gt": "goose", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15971 + }, + { + "path": "14141.flac", + "question": "What's driving in the video?", + "choice_a": "highway", + "choice_b": "bus", + "choice_c": "high-speed rail", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15972 + }, + { + "path": "49545.flac", + "question": "What happened in the video?", + "choice_a": "It's windy", + "choice_b": "Driving a car", + "choice_c": "Helicopter take off", + "choice_d": "Kites take off", + "answer_gt": "Helicopter take off", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15973 + }, + { + "path": "18865.flac", + "question": "What animal is in the video?", + "choice_a": "kitty", + "choice_b": "peacock", + "choice_c": "turkey", + "choice_d": "duck", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15974 + }, + { + "path": "50615.flac", + "question": "What animal appears in the video?", + "choice_a": "Seagull", + "choice_b": "cormorant", + "choice_c": "penguin", + "choice_d": "Big black bear", + "answer_gt": "penguin", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15975 + }, + { + "path": "8633.flac", + "question": "What is the main sound source of the video?", + "choice_a": "police car siren", + "choice_b": "Remote control car", + "choice_c": "Remote control ship", + "choice_d": "remote controlled aircraft", + "answer_gt": "remote controlled aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15976 + }, + { + "path": "27911.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Smoke Alarm", + "choice_b": "Fire engine siren", + "choice_c": "Ambulance siren", + "choice_d": "The train whistled", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15977 + }, + { + "path": "13323.flac", + "question": "What happened in the video?", + "choice_a": "fire", + "choice_b": "Set off firecrackers", + "choice_c": "firecrackers", + "choice_d": "Watering", + "answer_gt": "fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15978 + }, + { + "path": "48055.flac", + "question": "What happened in the video?", + "choice_a": "enjoy a hot spring", + "choice_b": "Water splashing", + "choice_c": "rowing", + "choice_d": "surfing", + "answer_gt": "Water splashing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15979 + }, + { + "path": "2851.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15980 + }, + { + "path": "24232.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "honeybee", + "choice_b": "cricket", + "choice_c": "grasshopper", + "choice_d": "mantis", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15981 + }, + { + "path": "40708.flac", + "question": "What animal appears in the video?", + "choice_a": "dove", + "choice_b": "eagle", + "choice_c": "owl", + "choice_d": "dove", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15982 + }, + { + "path": "52456.flac", + "question": "What animal appears in the video?", + "choice_a": "Magpie", + "choice_b": "go fishing", + "choice_c": "kangaroo", + "choice_d": "Woodpecker", + "answer_gt": "Woodpecker", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15983 + }, + { + "path": "21822.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "fox", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15984 + }, + { + "path": "21323.flac", + "question": "What animal is in the video?", + "choice_a": "owl", + "choice_b": "parrot", + "choice_c": "eagle", + "choice_d": "wild goose", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15985 + }, + { + "path": "57840.flac", + "question": "What are the people doing in the video?", + "choice_a": "physical exercise", + "choice_b": "practise qigong", + "choice_c": "Simulated swimming", + "choice_d": "Practicing Taiji", + "answer_gt": "Simulated swimming", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15986 + }, + { + "path": "15782.flac", + "question": "What animal appears in the video?", + "choice_a": "rowboat", + "choice_b": "Seals", + "choice_c": "whale", + "choice_d": "crocodile", + "answer_gt": "crocodile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15987 + }, + { + "path": "27859.flac", + "question": "Where did the video take place?", + "choice_a": "playground", + "choice_b": "pasture", + "choice_c": "Playground", + "choice_d": "zoo", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15988 + }, + { + "path": "15804.flac", + "question": "What are the people doing in the video?", + "choice_a": "performing oral skills", + "choice_b": "Machine gun fire", + "choice_c": "Beat birds", + "choice_d": "drive a nail", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15989 + }, + { + "path": "43088.flac", + "question": "Where does the video take place?", + "choice_a": "toilet", + "choice_b": "Amusement Park", + "choice_c": "zoo", + "choice_d": "farm", + "answer_gt": "farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15990 + }, + { + "path": "6552.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "sheep", + "choice_c": "horse", + "choice_d": "donkey", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15991 + }, + { + "path": "16626.flac", + "question": "What happened in the video?", + "choice_a": "snow", + "choice_b": "It's Hailing", + "choice_c": "rain", + "choice_d": "watering", + "answer_gt": "It's Hailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15992 + }, + { + "path": "61482.flac", + "question": "What is shown in the video?", + "choice_a": "Electric toothbrush", + "choice_b": "Electric hairdresser", + "choice_c": "Electric shaver", + "choice_d": "Electric face thinner", + "answer_gt": "Electric shaver", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15993 + }, + { + "path": "9314.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "vacuum cleaner to clean the floor", + "choice_c": "The sound of the stream", + "choice_d": "waterfall", + "answer_gt": "waterfall", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15994 + }, + { + "path": "42368.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorcycle", + "choice_b": "motorboat", + "choice_c": "truck", + "choice_d": "a storage battery car", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15995 + }, + { + "path": "1143.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "metro", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15996 + }, + { + "path": "3912.flac", + "question": "What animal is in the video?", + "choice_a": "chicken", + "choice_b": "dove", + "choice_c": "owl", + "choice_d": "duck", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15997 + }, + { + "path": "50375.flac", + "question": "What are the people in the video doing?", + "choice_a": "winter outdoor swimming", + "choice_b": "skiing", + "choice_c": "surfing", + "choice_d": "skating", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15998 + }, + { + "path": "16387.flac", + "question": "Where did the video take place?", + "choice_a": "Amusement Park", + "choice_b": "Skating rink", + "choice_c": "Ranching", + "choice_d": "Ski Field", + "answer_gt": "Ski Field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 15999 + }, + { + "path": "40658.flac", + "question": "What animal appears in the video?", + "choice_a": "planing", + "choice_b": "honeybee", + "choice_c": "cicada", + "choice_d": "Chicken", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16000 + }, + { + "path": "4573.flac", + "question": "What car is in the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "bus", + "choice_d": "motorboat", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16001 + }, + { + "path": "31889.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Car horn", + "choice_b": "Fire engine siren", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16002 + }, + { + "path": "58983.flac", + "question": "What happened in the video?", + "choice_a": "running", + "choice_b": "Pull the car and skateboard", + "choice_c": "Running after the car", + "choice_d": "Carjacking", + "answer_gt": "Pull the car and skateboard", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16003 + }, + { + "path": "61221.flac", + "question": "What animal is in the video?", + "choice_a": "tit", + "choice_b": "dove", + "choice_c": "Oriole", + "choice_d": "Hummingbird", + "answer_gt": "tit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16004 + }, + { + "path": "43772.flac", + "question": "Where does the video take place?", + "choice_a": "toilet", + "choice_b": "forest", + "choice_c": "zoo", + "choice_d": "farm", + "answer_gt": "farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16005 + }, + { + "path": "15432.flac", + "question": "What happened in the video?", + "choice_a": "watering", + "choice_b": "pump", + "choice_c": "The water pipe is leaking", + "choice_d": "Watering vegetable fields", + "answer_gt": "pump", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16006 + }, + { + "path": "3791.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "fire engine", + "choice_d": "helicopter", + "answer_gt": "fire engine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16007 + }, + { + "path": "60731.flac", + "question": "What are the people doing in the video?", + "choice_a": "speech", + "choice_b": "Performing oral skills", + "choice_c": "Performing and singing", + "choice_d": "Live oral skills course", + "answer_gt": "Performing oral skills", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16008 + }, + { + "path": "15165.flac", + "question": "What happened in the video?", + "choice_a": "Making toys", + "choice_b": "Knock something", + "choice_c": "Smash toys", + "choice_d": "Making bread", + "answer_gt": "Knock something", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16009 + }, + { + "path": "6599.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Toy car", + "choice_b": "Toy airplane", + "choice_c": "Toy motorcycle", + "choice_d": "Toy train", + "answer_gt": "Toy train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16010 + }, + { + "path": "31623.flac", + "question": "What are the people doing in the video?", + "choice_a": "roller coaster running", + "choice_b": "high jump", + "choice_c": "Rope skipping", + "choice_d": "dance", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16011 + }, + { + "path": "5172.flac", + "question": "What's driving in the video?", + "choice_a": "highway", + "choice_b": "Kayak", + "choice_c": "canoe", + "choice_d": "hovercraft", + "answer_gt": "canoe", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16012 + }, + { + "path": "10704.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "ambulance", + "choice_b": "motorcycle", + "choice_c": "motorboat", + "choice_d": "lawn mower", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16013 + }, + { + "path": "6378.flac", + "question": "What animal is in the video?", + "choice_a": "cattle", + "choice_b": "sheep", + "choice_c": "horse", + "choice_d": "donkey", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16014 + }, + { + "path": "36117.flac", + "question": "Where does the video take place?", + "choice_a": "waterfall", + "choice_b": "at sea", + "choice_c": "seabed", + "choice_d": "playground", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16015 + }, + { + "path": "55191.flac", + "question": "What is the main sound source of the video?", + "choice_a": "fire engine", + "choice_b": "The sound of the bus", + "choice_c": "The sound of ice cream cart", + "choice_d": "The sound of ice shaving", + "answer_gt": "The sound of ice cream cart", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16016 + }, + { + "path": "34975.flac", + "question": "What animal makes a sound in the video?", + "choice_a": "sheep", + "choice_b": "dog", + "choice_c": "fox", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16017 + }, + { + "path": "33496.flac", + "question": "What are the people doing in the video?", + "choice_a": "sing", + "choice_b": "Explain rope skipping", + "choice_c": "dance", + "choice_d": "perform", + "answer_gt": "Explain rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16018 + }, + { + "path": "44330.flac", + "question": "What are the people doing in the video?", + "choice_a": "take a shower", + "choice_b": "surfing", + "choice_c": "Diving", + "choice_d": "go fishing", + "answer_gt": "go fishing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16019 + }, + { + "path": "25663.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "goose", + "choice_c": "dog", + "choice_d": "duck", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16020 + }, + { + "path": "58883.flac", + "question": "What's on the toilet in this video?", + "choice_a": "Teddy bear", + "choice_b": "puppet", + "choice_c": "Big white rabbit", + "choice_d": "Rag doll", + "answer_gt": "Rag doll", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16021 + }, + { + "path": "7887.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "train", + "choice_c": "machine sound", + "choice_d": "vacuum cleaner to clean the floor", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16022 + }, + { + "path": "31258.flac", + "question": "Where did the video take place?", + "choice_a": "Swimming Pool", + "choice_b": "Grassland", + "choice_c": "sandy beach", + "choice_d": "at sea", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16023 + }, + { + "path": "24929.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "wolf", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16024 + }, + { + "path": "20339.flac", + "question": "What are the people in the video doing?", + "choice_a": "skiing", + "choice_b": "Killing chickens", + "choice_c": "Using a sewing machine", + "choice_d": "Feed the duck", + "answer_gt": "Using a sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16025 + }, + { + "path": "7104.flac", + "question": "What animal is in the video?", + "choice_a": "lion", + "choice_b": "wolf", + "choice_c": "elephant", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16026 + }, + { + "path": "3049.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sea wave", + "choice_b": "engine", + "choice_c": "Bubbling sound", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16027 + }, + { + "path": "61266.flac", + "question": "What happened in the video?", + "choice_a": "Tractor ploughing", + "choice_b": "Earth digging by excavator", + "choice_c": "Practice diving", + "choice_d": "Tractor digging", + "answer_gt": "Tractor ploughing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16028 + }, + { + "path": "4819.flac", + "question": "What are the people doing in the video?", + "choice_a": "staple books", + "choice_b": "sew clothes", + "choice_c": "typing", + "choice_d": "Make pants", + "answer_gt": "sew clothes", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16029 + }, + { + "path": "5901.flac", + "question": "What animal is in the video?", + "choice_a": "cheetah", + "choice_b": "otter", + "choice_c": "zebra", + "choice_d": "lion", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16030 + }, + { + "path": "6603.flac", + "question": "Where did the video take place?", + "choice_a": "aquatic", + "choice_b": "mountain path", + "choice_c": "field", + "choice_d": "highway", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16031 + }, + { + "path": "15507.flac", + "question": "What are the people in the video doing?", + "choice_a": "motorcycle", + "choice_b": "Mountaineering", + "choice_c": "skiing", + "choice_d": "typing on a typewriter", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16032 + }, + { + "path": "25492.flac", + "question": "What is the main sound source of the video?", + "choice_a": "ambulance", + "choice_b": "thunder", + "choice_c": "Civil defense alarm", + "choice_d": "Smoke Alarm", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16033 + }, + { + "path": "10768.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "it's hailing", + "choice_b": "wind", + "choice_c": "rain", + "choice_d": "Thundering", + "answer_gt": "wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16034 + }, + { + "path": "27984.flac", + "question": "Where did the video take place?", + "choice_a": "seabed", + "choice_b": "road", + "choice_c": "In the pond", + "choice_d": "beach", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16035 + }, + { + "path": "23092.flac", + "question": "What is the main sound source of the video?", + "choice_a": "motorcycle", + "choice_b": "bus", + "choice_c": "Tire burst", + "choice_d": "racing", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16036 + }, + { + "path": "47811.flac", + "question": "What happened in the video?", + "choice_a": "tornado", + "choice_b": "Volcanic explosion", + "choice_c": "tsunami", + "choice_d": "sand storm", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16037 + }, + { + "path": "40389.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sheep", + "choice_b": "mule", + "choice_c": "donkey", + "choice_d": "horse", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16038 + }, + { + "path": "5897.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "turkey", + "choice_b": "lion", + "choice_c": "Gibbon howls", + "choice_d": "kangaroo", + "answer_gt": "Gibbon howls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16039 + }, + { + "path": "57967.flac", + "question": "Where did the video take place?", + "choice_a": "farm", + "choice_b": "park", + "choice_c": "Grassland", + "choice_d": "On the road", + "answer_gt": "On the road", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16040 + }, + { + "path": "38012.flac", + "question": "What are the people doing in the video?", + "choice_a": "Shooting", + "choice_b": "Set off firecrackers", + "choice_c": "to shave", + "choice_d": "Beat birds", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16041 + }, + { + "path": "50299.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "The dog howled", + "choice_b": "Sheep crow", + "choice_c": "The wolf barked", + "choice_d": "The mouse barked", + "answer_gt": "The dog howled", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16042 + }, + { + "path": "57933.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Motorcycle engine noise", + "choice_b": "Clapping sound", + "choice_c": "Car engine noise", + "choice_d": "Lift canvas", + "answer_gt": "Motorcycle engine noise", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16043 + }, + { + "path": "59551.flac", + "question": "What is shown in the video?", + "choice_a": "The door is broken", + "choice_b": "I fell", + "choice_c": "The cupboard door is crooked", + "choice_d": "The window is broken", + "answer_gt": "The cupboard door is crooked", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16044 + }, + { + "path": "13566.flac", + "question": "What are the people in the video doing?", + "choice_a": "phone", + "choice_b": "Waiting for the bus", + "choice_c": "Get ready to cross the road", + "choice_d": "wait forsomeone", + "answer_gt": "Get ready to cross the road", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16045 + }, + { + "path": "51549.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The sound of rain", + "choice_b": "Airplane noise", + "choice_c": "The sound of whistle", + "choice_d": "The touch between the train and the track", + "answer_gt": "The touch between the train and the track", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16046 + }, + { + "path": "22117.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "firecrackers", + "choice_b": "tornado", + "choice_c": "cars crush things", + "choice_d": "avalanche", + "answer_gt": "tornado", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16047 + }, + { + "path": "53919.flac", + "question": "What is the main sound source in this video?", + "choice_a": "sound of wind", + "choice_b": "The sound of a horse's hoof", + "choice_c": "The crack of ice", + "choice_d": "The sound of stepping on the snow", + "answer_gt": "The sound of stepping on the snow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16048 + }, + { + "path": "51013.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "A rooster crows", + "choice_b": "phoenix", + "choice_c": "wild duck", + "choice_d": "Seabirds", + "answer_gt": "A rooster crows", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16049 + }, + { + "path": "15104.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "A rooster crows", + "choice_b": "fireworks and firecrackers", + "choice_c": "quack of a duck", + "choice_d": "Birdsong", + "answer_gt": "Birdsong", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16050 + }, + { + "path": "21028.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Fire engine siren", + "choice_b": "Civil defense alarm", + "choice_c": "bus", + "choice_d": "Ambulance siren", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16051 + }, + { + "path": "16770.flac", + "question": "What animal is in the video?", + "choice_a": "kitty", + "choice_b": "chicken", + "choice_c": "bird", + "choice_d": "Crab", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16052 + }, + { + "path": "25222.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "chicken", + "choice_b": "bird", + "choice_c": "duck", + "choice_d": "squirrel", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16053 + }, + { + "path": "1870.flac", + "question": "What kind of transportation does the person in the video drive?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "motorcycle", + "choice_d": "police car", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16054 + }, + { + "path": "25769.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cock", + "choice_c": "cat", + "choice_d": "peacock", + "answer_gt": "cock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16055 + }, + { + "path": "60522.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Sound of motorcycle running", + "choice_b": "The sound of ship engine", + "choice_c": "The sound of car driving", + "choice_d": "Truck exhaust pipe", + "answer_gt": "The sound of car driving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16056 + }, + { + "path": "50173.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Sheep crow", + "choice_b": "The wolf barked", + "choice_c": "The centipede barks", + "choice_d": "dog's bark", + "answer_gt": "dog's bark", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16057 + }, + { + "path": "2773.flac", + "question": "What are the people doing in the video?", + "choice_a": "Rope skipping", + "choice_b": "racing", + "choice_c": "skiing", + "choice_d": "rowing", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16058 + }, + { + "path": "27488.flac", + "question": "What animal is in the video?", + "choice_a": "cricket", + "choice_b": "Hamster", + "choice_c": "Vole", + "choice_d": "Otter", + "answer_gt": "Otter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16059 + }, + { + "path": "5788.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "ship", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16060 + }, + { + "path": "52723.flac", + "question": "What are the people in the video doing?", + "choice_a": "Skateboarding", + "choice_b": "Play balance car", + "choice_c": "Play Snowboarding", + "choice_d": "Play skating", + "answer_gt": "Skateboarding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16061 + }, + { + "path": "28371.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Underwater sound", + "choice_b": "sound of wind", + "choice_c": "thunder", + "choice_d": "The sound of the stream", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16062 + }, + { + "path": "27612.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "thunder", + "choice_b": "sound of wind", + "choice_c": "car engine noise", + "choice_d": "bomb", + "answer_gt": "thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16063 + }, + { + "path": "29903.flac", + "question": "What animal appears in the video?", + "choice_a": "cricket", + "choice_b": "Sea lions", + "choice_c": "hedgehog", + "choice_d": "Otter", + "answer_gt": "Otter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16064 + }, + { + "path": "18697.flac", + "question": "What animal appears in the video?", + "choice_a": "eagle", + "choice_b": "rabbit", + "choice_c": "owl", + "choice_d": "turtledove", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16065 + }, + { + "path": "16073.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "car engine noise", + "choice_b": "sound of wind", + "choice_c": "machine sound", + "choice_d": "vacuum cleaner to clean the floor", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16066 + }, + { + "path": "22396.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "horn", + "choice_b": "sound", + "choice_c": "horn", + "choice_d": "sound of wind", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16067 + }, + { + "path": "28723.flac", + "question": "What is the main sound source of the video?", + "choice_a": "lathe rotation sound", + "choice_b": "motorcycle", + "choice_c": "bus", + "choice_d": "metro", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16068 + }, + { + "path": "1235.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The sound of magpies", + "choice_b": "sounds made by chickens", + "choice_c": "The sound of ducks", + "choice_d": "Crows", + "answer_gt": "Crows", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16069 + }, + { + "path": "25304.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "goose", + "choice_c": "donkey", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16070 + }, + { + "path": "41217.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "aircraft", + "choice_c": "motorcycle", + "choice_d": "the sound of driving a snowmobile", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16071 + }, + { + "path": "37382.flac", + "question": "What animal appears in the video?", + "choice_a": "cheetah", + "choice_b": "black bear", + "choice_c": "lion", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16072 + }, + { + "path": "32347.flac", + "question": "What are the people doing in the video?", + "choice_a": "sow", + "choice_b": "drive a car", + "choice_c": "plough fields", + "choice_d": "cut grass", + "answer_gt": "cut grass", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16073 + }, + { + "path": "19928.flac", + "question": "What happened in the video?", + "choice_a": "Set off firecrackers", + "choice_b": "Setting off fireworks", + "choice_c": "blast", + "choice_d": "Smoke bomb", + "answer_gt": "Set off firecrackers", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16074 + }, + { + "path": "10661.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "motorcycle", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16075 + }, + { + "path": "20623.flac", + "question": "Where does the video take place?", + "choice_a": "seabed", + "choice_b": "highway", + "choice_c": "metro station", + "choice_d": "cage", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16076 + }, + { + "path": "28705.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "owl", + "choice_c": "crow", + "choice_d": "eagle", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16077 + }, + { + "path": "10996.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16078 + }, + { + "path": "54400.flac", + "question": "What happened to the video?", + "choice_a": "play tennis", + "choice_b": "shuttlecock", + "choice_c": "Play basketball", + "choice_d": "play table tennis", + "answer_gt": "shuttlecock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16079 + }, + { + "path": "32247.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16080 + }, + { + "path": "20859.flac", + "question": "What is the main sound source of the video?", + "choice_a": "car engine noise", + "choice_b": "hail", + "choice_c": "acceleration sound of automobile engine", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16081 + }, + { + "path": "8857.flac", + "question": "What animal is in the video?", + "choice_a": "owl", + "choice_b": "myna", + "choice_c": "dove", + "choice_d": "bird", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16082 + }, + { + "path": "54789.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Footsteps", + "choice_b": "The sound of pinching things", + "choice_c": "Tearing paper", + "choice_d": "The sound of stepping on the snow", + "answer_gt": "The sound of pinching things", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16083 + }, + { + "path": "9887.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "motorcycle", + "choice_b": "ambulance", + "choice_c": "racing", + "choice_d": "motorboat", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16084 + }, + { + "path": "36378.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Fireworks", + "choice_b": "thunder", + "choice_c": "shot", + "choice_d": "play the drum", + "answer_gt": "Fireworks", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16085 + }, + { + "path": "44620.flac", + "question": "Where did the video take place?", + "choice_a": "toilet", + "choice_b": "forest", + "choice_c": "in the air", + "choice_d": "grassland", + "answer_gt": "in the air", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16086 + }, + { + "path": "49759.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "helicopter", + "choice_b": "truck", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16087 + }, + { + "path": "48840.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Ambulance siren", + "choice_b": "horn", + "choice_c": "Bus horn", + "choice_d": "Truck horn", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16088 + }, + { + "path": "3088.flac", + "question": "What is the main sound source of the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "motorcycle", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16089 + }, + { + "path": "28844.flac", + "question": "What are the people doing in the video?", + "choice_a": "Swimming", + "choice_b": "go fishing", + "choice_c": "dance", + "choice_d": "crawl", + "answer_gt": "Swimming", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16090 + }, + { + "path": "20785.flac", + "question": "What animal appears in the video?", + "choice_a": "lion", + "choice_b": "elephant", + "choice_c": "cheetah", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16091 + }, + { + "path": "30543.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "shot", + "choice_c": "hail", + "choice_d": "rope skipping", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16092 + }, + { + "path": "49472.flac", + "question": "Where is the bird in the video?", + "choice_a": "aquatic", + "choice_b": "In the birdcage", + "choice_c": "In the grass", + "choice_d": "on the tree", + "answer_gt": "In the birdcage", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16093 + }, + { + "path": "43682.flac", + "question": "What animal appears in the video?", + "choice_a": "pig", + "choice_b": "monkey", + "choice_c": "Hyenas", + "choice_d": "hedgehog", + "answer_gt": "Hyenas", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16094 + }, + { + "path": "5763.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "chat", + "choice_b": "quarrel", + "choice_c": "Singing", + "choice_d": "People talk", + "answer_gt": "People talk", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16095 + }, + { + "path": "44703.flac", + "question": "Where did the video take place?", + "choice_a": "bus", + "choice_b": "Tunnel", + "choice_c": "highway", + "choice_d": "mountain path", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16096 + }, + { + "path": "27345.flac", + "question": "What are the people doing in the video?", + "choice_a": "rowing", + "choice_b": "Swimming", + "choice_c": "surfing", + "choice_d": "diving", + "answer_gt": "rowing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16097 + }, + { + "path": "45979.flac", + "question": "What animals appear in the video?", + "choice_a": "planing", + "choice_b": "long-horned grasshopper", + "choice_c": "mantis", + "choice_d": "grasshopper", + "answer_gt": "long-horned grasshopper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16098 + }, + { + "path": "61085.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Welding sound", + "choice_b": "Polish wood", + "choice_c": "The sound of planing", + "choice_d": "Polishing sound of jade", + "answer_gt": "The sound of planing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16099 + }, + { + "path": "58939.flac", + "question": "What is shown in the video?", + "choice_a": "Toilet flush", + "choice_b": "The toilet is clogged.", + "choice_c": "Sewer flushing", + "choice_d": "The sound of car wheels turning", + "answer_gt": "Toilet flush", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16100 + }, + { + "path": "11307.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "aircraft", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16101 + }, + { + "path": "48535.flac", + "question": "What happened in the video?", + "choice_a": "Volcano eruption", + "choice_b": "blast", + "choice_c": "tornado", + "choice_d": "volcanic explosion", + "answer_gt": "tornado", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16102 + }, + { + "path": "8429.flac", + "question": "What animal is in the video?", + "choice_a": "frog", + "choice_b": "dog", + "choice_c": "owl", + "choice_d": "turkey", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16103 + }, + { + "path": "24506.flac", + "question": "Where did the video take place?", + "choice_a": "aquatic", + "choice_b": "seabed", + "choice_c": "ski field", + "choice_d": "at sea", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16104 + }, + { + "path": "4234.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "duck", + "choice_c": "chicken", + "choice_d": "peacock", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16105 + }, + { + "path": "48067.flac", + "question": "Where did the video take place?", + "choice_a": "Underwater", + "choice_b": "On the sea", + "choice_c": "pond", + "choice_d": "By the beach", + "answer_gt": "Underwater", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16106 + }, + { + "path": "26021.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "rain", + "choice_c": "The sound of a waterfall", + "choice_d": "thunder", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16107 + }, + { + "path": "4615.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "highway", + "choice_d": "automobile", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16108 + }, + { + "path": "16184.flac", + "question": "What is the main sound source of the video?", + "choice_a": "ambulance", + "choice_b": "Civil defense alarm", + "choice_c": "Police car siren", + "choice_d": "The train whistled", + "answer_gt": "Police car siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16109 + }, + { + "path": "60614.flac", + "question": "What are the people doing in the video?", + "choice_a": "Competition Cuju", + "choice_b": "Play football in the game", + "choice_c": "Match boxing", + "choice_d": "Play volleyball in the match", + "answer_gt": "Play volleyball in the match", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16110 + }, + { + "path": "6906.flac", + "question": "What are the people doing in this video?", + "choice_a": "machine", + "choice_b": "live broadcast", + "choice_c": "speak", + "choice_d": "skiing", + "answer_gt": "speak", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16111 + }, + { + "path": "59015.flac", + "question": "What is shown in the video?", + "choice_a": "Dragon Boat Race", + "choice_b": "Cargo ship sailing", + "choice_c": "Sailing", + "choice_d": "Wandering on bamboo rafts", + "answer_gt": "Sailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16112 + }, + { + "path": "26953.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of wind", + "choice_b": "The sound of a waterfall", + "choice_c": "Underwater sound", + "choice_d": "thunder", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16113 + }, + { + "path": "56214.flac", + "question": "Where does the video take place?", + "choice_a": "On the road", + "choice_b": "On the grass", + "choice_c": "Beside the highway", + "choice_d": "Mud pit", + "answer_gt": "Beside the highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16114 + }, + { + "path": "27124.flac", + "question": "Where does the video take place?", + "choice_a": "playground", + "choice_b": "seabed", + "choice_c": "On land", + "choice_d": "racetrack", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16115 + }, + { + "path": "27129.flac", + "question": "What are the people doing in the video?", + "choice_a": "shot", + "choice_b": "Shooting", + "choice_c": "performing oral skills", + "choice_d": "rope skipping", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16116 + }, + { + "path": "24330.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "duck", + "choice_c": "bird", + "choice_d": "dog", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16117 + }, + { + "path": "9402.flac", + "question": "What are the people doing in the video?", + "choice_a": "skiing", + "choice_b": "racing", + "choice_c": "typing on a typewriter", + "choice_d": "play with water", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16118 + }, + { + "path": "54084.flac", + "question": "What animal is in the video?", + "choice_a": "monkey", + "choice_b": "Piggy", + "choice_c": "lamb", + "choice_d": "Calf", + "answer_gt": "Piggy", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16119 + }, + { + "path": "12378.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Driving a motorcycle", + "choice_b": "drive a car", + "choice_c": "aircraft", + "choice_d": "Repair the car", + "answer_gt": "Driving a motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16120 + }, + { + "path": "6718.flac", + "question": "What animal appears in the video?", + "choice_a": "crow", + "choice_b": "owl", + "choice_c": "dove", + "choice_d": "parrot", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16121 + }, + { + "path": "28595.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "truck", + "choice_c": "metro", + "choice_d": "bus", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16122 + }, + { + "path": "14097.flac", + "question": "What animal is in the video?", + "choice_a": "frog", + "choice_b": "duck", + "choice_c": "ostrich", + "choice_d": "chicken", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16123 + }, + { + "path": "24134.flac", + "question": "Where did the video take place?", + "choice_a": "seabed", + "choice_b": "seaside", + "choice_c": "desert", + "choice_d": "Grassland", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16124 + }, + { + "path": "47101.flac", + "question": "What are the people in the video doing?", + "choice_a": "to shave", + "choice_b": "lay an egg", + "choice_c": "Machine gun fire", + "choice_d": "Design", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16125 + }, + { + "path": "19503.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "metro", + "choice_b": "highway", + "choice_c": "bus", + "choice_d": "bus", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16126 + }, + { + "path": "34931.flac", + "question": "Where did the video take place?", + "choice_a": "at sea", + "choice_b": "zoo", + "choice_c": "field", + "choice_d": "suburb", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16127 + }, + { + "path": "29540.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sheep", + "choice_b": "horse", + "choice_c": "cattle", + "choice_d": "donkey", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16128 + }, + { + "path": "53013.flac", + "question": "What happened in the video?", + "choice_a": "Copying documents", + "choice_b": "Shredding documents", + "choice_c": "Printer paper jam", + "choice_d": "Printer printing", + "answer_gt": "Printer printing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16129 + }, + { + "path": "35923.flac", + "question": "What is the main sound source of the video?", + "choice_a": "automobile", + "choice_b": "motorcycle", + "choice_c": "lawn mower", + "choice_d": "electric saw", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16130 + }, + { + "path": "59948.flac", + "question": "What are the people doing in the video?", + "choice_a": "Performing oral skills", + "choice_b": "Performing and singing", + "choice_c": "Impromptu rap", + "choice_d": "have a stomachache", + "answer_gt": "Performing oral skills", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16131 + }, + { + "path": "47329.flac", + "question": "What are the people in the video doing?", + "choice_a": "rowing", + "choice_b": "Swimming", + "choice_c": "go fishing", + "choice_d": "match", + "answer_gt": "rowing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16132 + }, + { + "path": "15563.flac", + "question": "What animal appears in the video?", + "choice_a": "Magpie", + "choice_b": "swallow", + "choice_c": "myna", + "choice_d": "crow", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16133 + }, + { + "path": "6077.flac", + "question": "What are the people doing in the video?", + "choice_a": "Adjustable hoe", + "choice_b": "weed", + "choice_c": "seeder", + "choice_d": "Driving a motorcycle", + "answer_gt": "Adjustable hoe", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16134 + }, + { + "path": "46619.flac", + "question": "What are the people in the video doing?", + "choice_a": "go fishing", + "choice_b": "surfing", + "choice_c": "Archery", + "choice_d": "rowing", + "answer_gt": "rowing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16135 + }, + { + "path": "18690.flac", + "question": "Where did the video take place?", + "choice_a": "airport", + "choice_b": "Bus stop", + "choice_c": "wharf", + "choice_d": "Metro Station", + "answer_gt": "Metro Station", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16136 + }, + { + "path": "47640.flac", + "question": "Where did the video take place?", + "choice_a": "toilet", + "choice_b": "pasture", + "choice_c": "woods", + "choice_d": "field", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16137 + }, + { + "path": "2837.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "sheep", + "choice_c": "donkey", + "choice_d": "dog", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16138 + }, + { + "path": "46723.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "duck", + "choice_b": "dove", + "choice_c": "elephant", + "choice_d": "oriole", + "answer_gt": "duck", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16139 + }, + { + "path": "59484.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Car acceleration sound", + "choice_b": "Car horn", + "choice_c": "Motorcycle acceleration sound", + "choice_d": "Speed up of ATV", + "answer_gt": "Car acceleration sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16140 + }, + { + "path": "18617.flac", + "question": "What animal appears in the video?", + "choice_a": "kitty", + "choice_b": "turkey", + "choice_c": "ostrich", + "choice_d": "rabbit", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16141 + }, + { + "path": "37581.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "honeybee", + "choice_c": "eagle", + "choice_d": "dove", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16142 + }, + { + "path": "16206.flac", + "question": "What's the weather in the video?", + "choice_a": "a sunny day", + "choice_b": "cloudy", + "choice_c": "A snowy day", + "choice_d": "rainy day", + "answer_gt": "rainy day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16143 + }, + { + "path": "38018.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "lion", + "choice_c": "cheetah", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16144 + }, + { + "path": "7585.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "aircraft", + "choice_b": "rope skipping", + "choice_c": "tornado", + "choice_d": "Rain and thunder", + "answer_gt": "Rain and thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16145 + }, + { + "path": "32557.flac", + "question": "What is the main sound source of the video?", + "choice_a": "cattle", + "choice_b": "cricket", + "choice_c": "frog", + "choice_d": "cicada", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16146 + }, + { + "path": "28064.flac", + "question": "What are the people doing in the video?", + "choice_a": "Shooting", + "choice_b": "performing oral skills", + "choice_c": "Throwing darts", + "choice_d": "go hunting", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16147 + }, + { + "path": "29701.flac", + "question": "What sounds in the video?", + "choice_a": "motorcycle", + "choice_b": "automobile", + "choice_c": "ship", + "choice_d": "bus", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16148 + }, + { + "path": "23327.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "fire engine", + "choice_b": "aircraft", + "choice_c": "police car", + "choice_d": "Tractor", + "answer_gt": "police car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16149 + }, + { + "path": "18424.flac", + "question": "What are the people doing in this video?", + "choice_a": "Making clothes", + "choice_b": "Using a dust collector", + "choice_c": "Using a sewing machine", + "choice_d": "Using the ironing machine", + "answer_gt": "Using a sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16150 + }, + { + "path": "11169.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "turkey", + "choice_c": "dog", + "choice_d": "lion", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16151 + }, + { + "path": "60020.flac", + "question": "What is shown in the video?", + "choice_a": "Excavator loading", + "choice_b": "Tractor repair", + "choice_c": "Earth lifting by crane", + "choice_d": "Motorcycle start", + "answer_gt": "Excavator loading", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16152 + }, + { + "path": "55562.flac", + "question": "What happened in the video?", + "choice_a": "Meteor shower", + "choice_b": "Volleyball practice room", + "choice_c": "Candle burning", + "choice_d": "Setting off fireworks", + "answer_gt": "Setting off fireworks", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16153 + }, + { + "path": "60413.flac", + "question": "What animal is in the video?", + "choice_a": "Woodpecker", + "choice_b": "Oriole", + "choice_c": "Hummingbird", + "choice_d": "Amber Parrot", + "answer_gt": "Woodpecker", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16154 + }, + { + "path": "60664.flac", + "question": "What happened in the video?", + "choice_a": "A dog bites a pig", + "choice_b": "The wolf bites the sheep", + "choice_c": "A dog bites a cow", + "choice_d": "Wolf bites rhinoceros", + "answer_gt": "A dog bites a pig", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16155 + }, + { + "path": "15846.flac", + "question": "What happened in the video?", + "choice_a": "mosquito", + "choice_b": "Watch the flies", + "choice_c": "Observe ants", + "choice_d": "Observe bees", + "answer_gt": "Observe bees", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16156 + }, + { + "path": "16226.flac", + "question": "What animal appears in the video?", + "choice_a": "cheetah", + "choice_b": "tiger", + "choice_c": "lion", + "choice_d": "The bear", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16157 + }, + { + "path": "14779.flac", + "question": "What are the people doing in the video?", + "choice_a": "riding", + "choice_b": "racing", + "choice_c": "skiing", + "choice_d": "hunting", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16158 + }, + { + "path": "55223.flac", + "question": "What are the people in the video doing?", + "choice_a": "Driving a car", + "choice_b": "By cable car", + "choice_c": "Driving a snowmobile", + "choice_d": "Sledding", + "answer_gt": "Driving a snowmobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16159 + }, + { + "path": "25046.flac", + "question": "Where does the video take place?", + "choice_a": "seabed", + "choice_b": "field", + "choice_c": "beach", + "choice_d": "at sea", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16160 + }, + { + "path": "48171.flac", + "question": "What season did the video take place?", + "choice_a": "winter", + "choice_b": "summer", + "choice_c": "autumn", + "choice_d": "spring", + "answer_gt": "winter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16161 + }, + { + "path": "14525.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "fox", + "choice_c": "wolf", + "choice_d": "cat", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16162 + }, + { + "path": "50293.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Car horn", + "choice_b": "honk", + "choice_c": "Fire engine siren", + "choice_d": "Truck horn", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16163 + }, + { + "path": "10261.flac", + "question": "What are the people doing in this video?", + "choice_a": "performing oral skills", + "choice_b": "drive a nail", + "choice_c": "Infrared induction", + "choice_d": "Gun fire", + "answer_gt": "Gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16164 + }, + { + "path": "5119.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "motorboat", + "choice_c": "motorcycle", + "choice_d": "Water flow sound", + "answer_gt": "motorboat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16165 + }, + { + "path": "19455.flac", + "question": "Where does the video take place?", + "choice_a": "at sea", + "choice_b": "By the pond", + "choice_c": "Xiaoyu", + "choice_d": "seaside", + "answer_gt": "seaside", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16166 + }, + { + "path": "4021.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "play the drum", + "choice_b": "Ambulance siren", + "choice_c": "The sound of music", + "choice_d": "Air defense alert", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16167 + }, + { + "path": "18373.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Fire engine siren", + "choice_b": "honk", + "choice_c": "Pre action", + "choice_d": "Ambulance siren", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16168 + }, + { + "path": "44584.flac", + "question": "Where did the video take place?", + "choice_a": "toilet", + "choice_b": "zoo", + "choice_c": "indoor", + "choice_d": "pet shop", + "answer_gt": "indoor", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16169 + }, + { + "path": "43777.flac", + "question": "Where does the video take place?", + "choice_a": "street", + "choice_b": "racetrack", + "choice_c": "In the cattle pen", + "choice_d": "field", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16170 + }, + { + "path": "37196.flac", + "question": "What animal appears in the video?", + "choice_a": "turkey", + "choice_b": "goose", + "choice_c": "peacock", + "choice_d": "vulture", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16171 + }, + { + "path": "42127.flac", + "question": "What animal appears in the video?", + "choice_a": "honeybee", + "choice_b": "planing", + "choice_c": "mosquito", + "choice_d": "snake", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16172 + }, + { + "path": "51242.flac", + "question": "Where is the dog in the video?", + "choice_a": "In the kennel", + "choice_b": "on the tree", + "choice_c": "In the basket", + "choice_d": "In the basket", + "answer_gt": "In the kennel", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16173 + }, + { + "path": "24462.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "dove", + "choice_c": "pheasant", + "choice_d": "peacock", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16174 + }, + { + "path": "19462.flac", + "question": "What is the person in the video holding?", + "choice_a": "sound of wind", + "choice_b": "rocket", + "choice_c": "Pistol", + "choice_d": "bowling", + "answer_gt": "Pistol", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16175 + }, + { + "path": "35196.flac", + "question": "What's driving in the video?", + "choice_a": "aircraft", + "choice_b": "metro", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16176 + }, + { + "path": "47742.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "Set off firecrackers", + "choice_c": "the purr of a cat", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16177 + }, + { + "path": "9257.flac", + "question": "What are the animals doing in the video?", + "choice_a": "sleep", + "choice_b": "skiing", + "choice_c": "fight", + "choice_d": "Competing for prey", + "answer_gt": "fight", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16178 + }, + { + "path": "50413.flac", + "question": "What are the people in the video doing?", + "choice_a": "Diving", + "choice_b": "surfing", + "choice_c": "in the mountains", + "choice_d": "Swimming", + "answer_gt": "Swimming", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16179 + }, + { + "path": "52359.flac", + "question": "What is shown in the video?", + "choice_a": "scenery", + "choice_b": "Vehicle technology", + "choice_c": "Parking lot", + "choice_d": "Building", + "answer_gt": "Vehicle technology", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16180 + }, + { + "path": "46665.flac", + "question": "What is shown in the video?", + "choice_a": "machine", + "choice_b": "grassland", + "choice_c": "fruit tree", + "choice_d": "gather apples", + "answer_gt": "fruit tree", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16181 + }, + { + "path": "14529.flac", + "question": "What's driving in the video?", + "choice_a": "motorcycle", + "choice_b": "train", + "choice_c": "highway", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16182 + }, + { + "path": "11080.flac", + "question": "What animal is in the video?", + "choice_a": "chicken", + "choice_b": "Petrel", + "choice_c": "cockcrow", + "choice_d": "eagle", + "answer_gt": "eagle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16183 + }, + { + "path": "5676.flac", + "question": "What are the people doing in the video?", + "choice_a": "drift", + "choice_b": "sailing", + "choice_c": "enjoy a hot spring", + "choice_d": "setting off fireworks", + "answer_gt": "drift", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16184 + }, + { + "path": "21256.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "sheep", + "choice_c": "cat", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16185 + }, + { + "path": "56719.flac", + "question": "Where does the video take place?", + "choice_a": "Toy car", + "choice_b": "In the field", + "choice_c": "Grassland", + "choice_d": "beach", + "answer_gt": "In the field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16186 + }, + { + "path": "16706.flac", + "question": "What are the people doing in the video?", + "choice_a": "A snowy day", + "choice_b": "Using the printer", + "choice_c": "Using a sewing machine", + "choice_d": "Use dust collector]", + "answer_gt": "Using a sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16187 + }, + { + "path": "41745.flac", + "question": "What animal is in the video?", + "choice_a": "planing", + "choice_b": "elephant", + "choice_c": "crocodile", + "choice_d": "frog", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16188 + }, + { + "path": "33639.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "ship", + "choice_c": "bus", + "choice_d": "motorcycle", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16189 + }, + { + "path": "46841.flac", + "question": "What happened in the video?", + "choice_a": "Reversing", + "choice_b": "Test Tire Quality", + "choice_c": "Clean the tire", + "choice_d": "Cars press things", + "answer_gt": "Cars press things", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16190 + }, + { + "path": "20415.flac", + "question": "What animal is in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16191 + }, + { + "path": "18124.flac", + "question": "What animal appears in the video?", + "choice_a": "Seagulls and chickens", + "choice_b": "Dog and rabbit", + "choice_c": "Swan and cat", + "choice_d": "Penguins and birds", + "answer_gt": "Penguins and birds", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16192 + }, + { + "path": "35147.flac", + "question": "What animal in the video makes a sound?", + "choice_a": "cat", + "choice_b": "sheep", + "choice_c": "dog", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16193 + }, + { + "path": "7287.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "horn", + "choice_c": "automobile", + "choice_d": "motorcycle", + "answer_gt": "horn", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16194 + }, + { + "path": "36002.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "sheep", + "choice_c": "elephant", + "choice_d": "pig", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16195 + }, + { + "path": "3223.flac", + "question": "What animal is in the video?", + "choice_a": "bird", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "dove", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16196 + }, + { + "path": "28895.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "chicken", + "choice_b": "cat", + "choice_c": "owl", + "choice_d": "eagle", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16197 + }, + { + "path": "8715.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "aircraft", + "choice_d": "ship", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16198 + }, + { + "path": "12737.flac", + "question": "What animal appears in the video?", + "choice_a": "parrot", + "choice_b": "peacock", + "choice_c": "pheasant", + "choice_d": "ostrich", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16199 + }, + { + "path": "24125.flac", + "question": "What is the main sound source of the video?", + "choice_a": "car engine noise", + "choice_b": "wind", + "choice_c": "Underwater sound", + "choice_d": "acceleration sound of automobile engine", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16200 + }, + { + "path": "40242.flac", + "question": "What happened in the video?", + "choice_a": "Volcano eruption", + "choice_b": "Setting off fireworks", + "choice_c": "volcanic explosion", + "choice_d": "fire", + "answer_gt": "fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16201 + }, + { + "path": "26701.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "chimpanzee", + "choice_c": "Gibbon", + "choice_d": "Golden Monkey", + "answer_gt": "Gibbon", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16202 + }, + { + "path": "31231.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "sheep", + "choice_d": "deer", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16203 + }, + { + "path": "59202.flac", + "question": "What is the main sound source in this video?", + "choice_a": "The sound of the trumpet", + "choice_b": "Welding sound", + "choice_c": "The sound of planing", + "choice_d": "Cutting sound", + "answer_gt": "The sound of planing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16204 + }, + { + "path": "42077.flac", + "question": "What are the people in the video doing?", + "choice_a": "roller coaster running", + "choice_b": "run", + "choice_c": "skating", + "choice_d": "skiing", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16205 + }, + { + "path": "57414.flac", + "question": "What happened in the video?", + "choice_a": "Group fight", + "choice_b": "Group dancing", + "choice_c": "There was a riot", + "choice_d": "Electric chair", + "answer_gt": "There was a riot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16206 + }, + { + "path": "36761.flac", + "question": "What are the people doing in the video?", + "choice_a": "skiing", + "choice_b": "run", + "choice_c": "Mountaineering", + "choice_d": "high jump", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16207 + }, + { + "path": "32907.flac", + "question": "What are the people in the video doing?", + "choice_a": "Cutting plastic bottles", + "choice_b": "paper-cut", + "choice_c": "flower arrangement", + "choice_d": "do homework", + "answer_gt": "Cutting plastic bottles", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16208 + }, + { + "path": "27308.flac", + "question": "Where does the video take place?", + "choice_a": "field", + "choice_b": "seaside", + "choice_c": "park", + "choice_d": "of one's own unit", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16209 + }, + { + "path": "16249.flac", + "question": "Where did the video take place?", + "choice_a": "street", + "choice_b": "in the water", + "choice_c": "aquatic", + "choice_d": "In the forest", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16210 + }, + { + "path": "30315.flac", + "question": "What animal appears in the video?", + "choice_a": "crow", + "choice_b": "goose", + "choice_c": "duck", + "choice_d": "pig", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16211 + }, + { + "path": "34384.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "seabed", + "choice_c": "at sea", + "choice_d": "Grassland", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16212 + }, + { + "path": "53270.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The plane is moving", + "choice_b": "exhaust pipe", + "choice_c": "automobile engine", + "choice_d": "Tractor running", + "answer_gt": "automobile engine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16213 + }, + { + "path": "16483.flac", + "question": "Where did the video take place?", + "choice_a": "airport", + "choice_b": "Bus stop", + "choice_c": "market", + "choice_d": "Metro Station", + "answer_gt": "Metro Station", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16214 + }, + { + "path": "26120.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16215 + }, + { + "path": "28544.flac", + "question": "What animal appears in the video?", + "choice_a": "duck", + "choice_b": "dog", + "choice_c": "wild goose", + "choice_d": "goose", + "answer_gt": "goose", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16216 + }, + { + "path": "34649.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "racing", + "choice_c": "motorcycle", + "choice_d": "gun fire", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16217 + }, + { + "path": "30737.flac", + "question": "What is the (main) sound source in the video?", + "choice_a": "sound of rain", + "choice_b": "It's windy", + "choice_c": "The sound of a waterfall", + "choice_d": "thunder", + "answer_gt": "thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16218 + }, + { + "path": "1206.flac", + "question": "What animal is in the video?", + "choice_a": "dove", + "choice_b": "chicken", + "choice_c": "parrot", + "choice_d": "Seagull", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16219 + }, + { + "path": "32980.flac", + "question": "Where does the video take place?", + "choice_a": "seaside", + "choice_b": "highway", + "choice_c": "seabed", + "choice_d": "park", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16220 + }, + { + "path": "49585.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Drums", + "choice_b": "cowbell", + "choice_c": "rope skipping", + "choice_d": "Airplane noise", + "answer_gt": "cowbell", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16221 + }, + { + "path": "32335.flac", + "question": "Where does the video take place?", + "choice_a": "On the subway", + "choice_b": "of one's own unit", + "choice_c": "On the plane", + "choice_d": "In the car", + "answer_gt": "In the car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16222 + }, + { + "path": "40411.flac", + "question": "What's the weather in the video?", + "choice_a": "rain", + "choice_b": "a sunny day", + "choice_c": "Foggy days", + "choice_d": "overcast", + "answer_gt": "a sunny day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16223 + }, + { + "path": "19844.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "acceleration sound of automobile engine", + "choice_c": "aircraft", + "choice_d": "gun", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16224 + }, + { + "path": "12526.flac", + "question": "Where did the video take place?", + "choice_a": "aquatic", + "choice_b": "train station", + "choice_c": "woods", + "choice_d": "the river front", + "answer_gt": "woods", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16225 + }, + { + "path": "3843.flac", + "question": "What animal is in the video?", + "choice_a": "frog", + "choice_b": "duck", + "choice_c": "penguin", + "choice_d": "monkey", + "answer_gt": "penguin", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16226 + }, + { + "path": "57458.flac", + "question": "What is the white on the ground in this video?", + "choice_a": "chemical fertilizer", + "choice_b": "White noodles", + "choice_c": "Snow white", + "choice_d": "lime", + "answer_gt": "Snow white", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16227 + }, + { + "path": "10647.flac", + "question": "What are the people doing in the video?", + "choice_a": "touch the cat", + "choice_b": "Catching earthworms", + "choice_c": "Snake catching", + "choice_d": "Release", + "answer_gt": "Snake catching", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16228 + }, + { + "path": "43826.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Birdsong", + "choice_b": "The peacock crows", + "choice_c": "Cockcrow", + "choice_d": "quack of a duck", + "answer_gt": "Cockcrow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16229 + }, + { + "path": "35773.flac", + "question": "Where did the video take place?", + "choice_a": "playground", + "choice_b": "at sea", + "choice_c": "aquatic", + "choice_d": "of one's own unit", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16230 + }, + { + "path": "39401.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "fox", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "fox", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16231 + }, + { + "path": "37347.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "automobile", + "choice_d": "metro", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16232 + }, + { + "path": "15343.flac", + "question": "Where did the video take place?", + "choice_a": "airport", + "choice_b": "Bus stop", + "choice_c": "wharf", + "choice_d": "Metro Station", + "answer_gt": "Metro Station", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16233 + }, + { + "path": "6382.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "aircraft", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16234 + }, + { + "path": "10533.flac", + "question": "Where did the video take place?", + "choice_a": "square", + "choice_b": "field", + "choice_c": "Playground", + "choice_d": "On the street", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16235 + }, + { + "path": "26779.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "pig", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16236 + }, + { + "path": "47201.flac", + "question": "What happened in the video?", + "choice_a": "Thunder", + "choice_b": "rain", + "choice_c": "The bear", + "choice_d": "hold an umbrella", + "answer_gt": "Thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16237 + }, + { + "path": "29483.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "racing", + "choice_b": "motorboat", + "choice_c": "fireworks and firecrackers", + "choice_d": "motorcycle", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16238 + }, + { + "path": "34662.flac", + "question": "What are the people doing in the video?", + "choice_a": "Shooting", + "choice_b": "to shave", + "choice_c": "rope skipping", + "choice_d": "Shooting", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16239 + }, + { + "path": "21002.flac", + "question": "What are the people in the video doing?", + "choice_a": "high jump", + "choice_b": "skating", + "choice_c": "dance", + "choice_d": "skiing", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16240 + }, + { + "path": "25744.flac", + "question": "What animal appears in the video?", + "choice_a": "bird", + "choice_b": "goose", + "choice_c": "dog", + "choice_d": "cat", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16241 + }, + { + "path": "56644.flac", + "question": "What are the people doing in the video?", + "choice_a": "Sharpen the knife", + "choice_b": "Cutting wood", + "choice_c": "Planing", + "choice_d": "polishing", + "answer_gt": "Planing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16242 + }, + { + "path": "20904.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "rowboat", + "choice_c": "whale", + "choice_d": "fish", + "answer_gt": "whale", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16243 + }, + { + "path": "32211.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sheep", + "choice_b": "cow-bells", + "choice_c": "The cat barks", + "choice_d": "organ", + "answer_gt": "cow-bells", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16244 + }, + { + "path": "45705.flac", + "question": "What is the sound source of the video?", + "choice_a": "firecrackers", + "choice_b": "Gun fire", + "choice_c": "crushing of plastic bottles", + "choice_d": "Beat the drum", + "answer_gt": "Gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16245 + }, + { + "path": "1854.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "Tractor", + "choice_c": "train", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16246 + }, + { + "path": "28270.flac", + "question": "What animal is in the video?", + "choice_a": "chicken", + "choice_b": "thrush", + "choice_c": "bird", + "choice_d": "e", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16247 + }, + { + "path": "56762.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Toilet flush", + "choice_b": "Sewer flushing", + "choice_c": "Motor pumping", + "choice_d": "Motor pumping sound", + "answer_gt": "Toilet flush", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16248 + }, + { + "path": "23256.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "fireworks and firecrackers", + "choice_c": "Bus", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16249 + }, + { + "path": "60251.flac", + "question": "What is shown in the video?", + "choice_a": "Weaving on loom", + "choice_b": "The photo is developed by the developing machine", + "choice_c": "Printer printing", + "choice_d": "Answer the fax", + "answer_gt": "Printer printing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16250 + }, + { + "path": "26065.flac", + "question": "What are the people in the video doing?", + "choice_a": "go fishing", + "choice_b": "Swimming", + "choice_c": "rowing", + "choice_d": "surfing", + "answer_gt": "go fishing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16251 + }, + { + "path": "27292.flac", + "question": "Where does the video take place?", + "choice_a": "Metro Station", + "choice_b": "mountain path", + "choice_c": "Tunnel", + "choice_d": "cave", + "answer_gt": "Tunnel", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16252 + }, + { + "path": "52149.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "The bear barked", + "choice_b": "The lion howled", + "choice_c": "Gibbon howls", + "choice_d": "Whistle", + "answer_gt": "Gibbon howls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16253 + }, + { + "path": "19428.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "seabed", + "choice_c": "ski field", + "choice_d": "On the beach", + "answer_gt": "seaside", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16254 + }, + { + "path": "51509.flac", + "question": "What happened in the video?", + "choice_a": "Mending shoes", + "choice_b": "Trimming trouser legs", + "choice_c": "Embroidery", + "choice_d": "Trim a coat", + "answer_gt": "Embroidery", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16255 + }, + { + "path": "8225.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "car engine noise", + "choice_b": "gun fire", + "choice_c": "Racing sound", + "choice_d": "sound of wind", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16256 + }, + { + "path": "18296.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Ambulance siren", + "choice_b": "Fire engine siren", + "choice_c": "Smoke Alarm", + "choice_d": "Civil defense alarm", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16257 + }, + { + "path": "50232.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "cricket", + "choice_b": "fly", + "choice_c": "tortoise", + "choice_d": "zebra", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16258 + }, + { + "path": "10986.flac", + "question": "What animal is in the video?", + "choice_a": "tortoise", + "choice_b": "crocodile", + "choice_c": "buffalo", + "choice_d": "Python", + "answer_gt": "crocodile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16259 + }, + { + "path": "40229.flac", + "question": "What animal is in the video?", + "choice_a": "turkey", + "choice_b": "goose", + "choice_c": "dove", + "choice_d": "bird", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16260 + }, + { + "path": "37804.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "dog", + "choice_c": "pig", + "choice_d": "chipmunks", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16261 + }, + { + "path": "60310.flac", + "question": "What is shown in the video?", + "choice_a": "Knitting sweater", + "choice_b": "Photo developing machine", + "choice_c": "Large printer", + "choice_d": "Large copier", + "answer_gt": "Large printer", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16262 + }, + { + "path": "60271.flac", + "question": "What animal is in the video?", + "choice_a": "Poodle", + "choice_b": "Kitty", + "choice_c": "Wild wolf", + "choice_d": "Native pig", + "answer_gt": "Poodle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16263 + }, + { + "path": "26202.flac", + "question": "Where did the video take place?", + "choice_a": "Wildlife Park", + "choice_b": "playground", + "choice_c": "ski field", + "choice_d": "Amusement Park", + "answer_gt": "Wildlife Park", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16264 + }, + { + "path": "3149.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "duck", + "choice_c": "peacock", + "choice_d": "chicken", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16265 + }, + { + "path": "47254.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "aircraft", + "choice_b": "train", + "choice_c": "bus", + "choice_d": "fuel tank car", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16266 + }, + { + "path": "19957.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Wind chime", + "choice_b": "siren", + "choice_c": "one-arm bandit", + "choice_d": "small bell", + "answer_gt": "small bell", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16267 + }, + { + "path": "9637.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "car engine noise", + "choice_c": "hail", + "choice_d": "Thundering", + "answer_gt": "Thundering", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16268 + }, + { + "path": "37902.flac", + "question": "Where did the video take place?", + "choice_a": "playground", + "choice_b": "Playground", + "choice_c": "zoo", + "choice_d": "On the street", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16269 + }, + { + "path": "43960.flac", + "question": "Where did the video take place?", + "choice_a": "elephant", + "choice_b": "in the mountains", + "choice_c": "grassland", + "choice_d": "woods", + "answer_gt": "woods", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16270 + }, + { + "path": "9317.flac", + "question": "What happened in the video?", + "choice_a": "tornado", + "choice_b": "pump", + "choice_c": "Volcanic explosion", + "choice_d": "volcanic explosion", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16271 + }, + { + "path": "36659.flac", + "question": "What animal is in the video?", + "choice_a": "cicada", + "choice_b": "cricket", + "choice_c": "grasshopper", + "choice_d": "mantis", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16272 + }, + { + "path": "46867.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "tortoise", + "choice_c": "Sea lions", + "choice_d": "crocodile", + "answer_gt": "tortoise", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16273 + }, + { + "path": "58087.flac", + "question": "What are the people doing in the video?", + "choice_a": "drink water", + "choice_b": "sing", + "choice_c": "listen to the music", + "choice_d": "Eat something", + "answer_gt": "Eat something", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16274 + }, + { + "path": "20098.flac", + "question": "What animal is in the video?", + "choice_a": "kitty", + "choice_b": "chicken", + "choice_c": "bird", + "choice_d": "duck", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16275 + }, + { + "path": "31730.flac", + "question": "What is the main sound source of the video?", + "choice_a": "motorboat", + "choice_b": "bus", + "choice_c": "high-speed rail", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16276 + }, + { + "path": "33659.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "aircraft", + "choice_b": "train", + "choice_c": "motorboat", + "choice_d": "the sound of driving a snowmobile", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16277 + }, + { + "path": "14233.flac", + "question": "What animal is in the video?", + "choice_a": "Seals", + "choice_b": "crocodile", + "choice_c": "tortoise", + "choice_d": "Sea lions", + "answer_gt": "crocodile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16278 + }, + { + "path": "58482.flac", + "question": "What are the people doing in the video?", + "choice_a": "play football", + "choice_b": "Bowling", + "choice_c": "practice oral skills", + "choice_d": "sharpen the knife", + "answer_gt": "Bowling", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16279 + }, + { + "path": "48800.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "thunder", + "choice_b": "The sound of snow", + "choice_c": "The sound of rain", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16280 + }, + { + "path": "13067.flac", + "question": "What are the people in the video doing?", + "choice_a": "Driving a motorcycle", + "choice_b": "honk", + "choice_c": "photograph", + "choice_d": "Repairing motorcycles", + "answer_gt": "honk", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16281 + }, + { + "path": "15183.flac", + "question": "What are the people doing in the video?", + "choice_a": "Cycling", + "choice_b": "fly an airplane", + "choice_c": "Driving a tractor", + "choice_d": "drive a car", + "answer_gt": "drive a car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16282 + }, + { + "path": "9065.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "chipmunks", + "choice_c": "bird", + "choice_d": "canary", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16283 + }, + { + "path": "38115.flac", + "question": "What animals appear in the video?", + "choice_a": "honeybee", + "choice_b": "dove", + "choice_c": "monkey", + "choice_d": "Canary", + "answer_gt": "Canary", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16284 + }, + { + "path": "18742.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Smoke Alarm", + "choice_b": "Civil defense alarm", + "choice_c": "flute", + "choice_d": "Siren", + "answer_gt": "Siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16285 + }, + { + "path": "47611.flac", + "question": "What is the sound source of the video?", + "choice_a": "sound of rain", + "choice_b": "The sound of rain", + "choice_c": "The sound of a waterfall", + "choice_d": "thunder", + "answer_gt": "The sound of a waterfall", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16286 + }, + { + "path": "22225.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "wolf", + "choice_c": "cat", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16287 + }, + { + "path": "41226.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "to shave", + "choice_b": "Shooting", + "choice_c": "Shooting", + "choice_d": "eagle", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16288 + }, + { + "path": "35143.flac", + "question": "What animal is in the video?", + "choice_a": "peacock", + "choice_b": "duck", + "choice_c": "chimpanzee", + "choice_d": "turkey", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16289 + }, + { + "path": "17561.flac", + "question": "What animal is in the video?", + "choice_a": "fox", + "choice_b": "chimpanzee", + "choice_c": "Gibbon", + "choice_d": "Golden Monkey", + "answer_gt": "Gibbon", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16290 + }, + { + "path": "25154.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "dog", + "choice_c": "cat", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16291 + }, + { + "path": "9149.flac", + "question": "What are the people doing in the video?", + "choice_a": "sailing", + "choice_b": "go fishing", + "choice_c": "rowing", + "choice_d": "flush the toilet", + "answer_gt": "go fishing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16292 + }, + { + "path": "32128.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Wind chime", + "choice_b": "Car horn", + "choice_c": "Police car siren", + "choice_d": "Train horn", + "answer_gt": "Police car siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16293 + }, + { + "path": "3262.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sheep", + "choice_b": "wolf", + "choice_c": "dog", + "choice_d": "Human voice", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16294 + }, + { + "path": "55136.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Doorbell", + "choice_b": "Piano sound", + "choice_c": "The bell rang", + "choice_d": "Mobile phone sound", + "answer_gt": "Mobile phone sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16295 + }, + { + "path": "14412.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "pig", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16296 + }, + { + "path": "16493.flac", + "question": "What are the people in the car doing in this video?", + "choice_a": "sleep", + "choice_b": "by car", + "choice_c": "photograph", + "choice_d": "wait forsomeone", + "answer_gt": "photograph", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16297 + }, + { + "path": "6687.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "chicken", + "choice_c": "bird", + "choice_d": "duck", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16298 + }, + { + "path": "4388.flac", + "question": "What is the main sound source of the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "motorcycle", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16299 + }, + { + "path": "59174.flac", + "question": "What is shown in the video?", + "choice_a": "Mop for water", + "choice_b": "Sweep the floor with a broom", + "choice_c": "Garbage robot mopping", + "choice_d": "Cleaning garbage with vacuum cleaner", + "answer_gt": "Cleaning garbage with vacuum cleaner", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16300 + }, + { + "path": "54010.flac", + "question": "Where did the video take place?", + "choice_a": "highway", + "choice_b": "stadium", + "choice_c": "Racetrack", + "choice_d": "On the track", + "answer_gt": "Racetrack", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16301 + }, + { + "path": "1985.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "dog", + "choice_c": "bird", + "choice_d": "kangaroo", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16302 + }, + { + "path": "38623.flac", + "question": "Where is the bird in the video?", + "choice_a": "On the grass", + "choice_b": "In the grass", + "choice_c": "In the birdcage", + "choice_d": "on the tree", + "answer_gt": "In the birdcage", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16303 + }, + { + "path": "25972.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "The sound of a waterfall", + "choice_c": "rain", + "choice_d": "thunder", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16304 + }, + { + "path": "49348.flac", + "question": "Where did the video take place?", + "choice_a": "square", + "choice_b": "Training ground", + "choice_c": "Racetrack", + "choice_d": "Airplane Lane", + "answer_gt": "Racetrack", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16305 + }, + { + "path": "58892.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Car horn", + "choice_b": "Automobile engine knock", + "choice_c": "Motorcycle engine knock", + "choice_d": "Wheel noise", + "answer_gt": "Automobile engine knock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16306 + }, + { + "path": "27703.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "dog", + "choice_c": "fox", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16307 + }, + { + "path": "29159.flac", + "question": "What animal makes a sound in the video?", + "choice_a": "planing", + "choice_b": "mosquito", + "choice_c": "honeybee", + "choice_d": "Ants", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16308 + }, + { + "path": "39162.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "duck", + "choice_c": "chicken", + "choice_d": "peacock", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16309 + }, + { + "path": "38807.flac", + "question": "What's the weather in the video?", + "choice_a": "thunderstorm", + "choice_b": "cloudy", + "choice_c": "downpour", + "choice_d": "a sunny day", + "answer_gt": "thunderstorm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16310 + }, + { + "path": "10132.flac", + "question": "What animal is in the video?", + "choice_a": "owl", + "choice_b": "duck", + "choice_c": "bird", + "choice_d": "ostrich", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16311 + }, + { + "path": "57594.flac", + "question": "What happened in the video?", + "choice_a": "Start the electric saw", + "choice_b": "There's an oil leak", + "choice_c": "tank", + "choice_d": "Automobile engine knock", + "answer_gt": "Automobile engine knock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16312 + }, + { + "path": "20325.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "frog", + "choice_c": "Toad", + "choice_d": "Giant salamander", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16313 + }, + { + "path": "3506.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16314 + }, + { + "path": "24618.flac", + "question": "What is the main sound source of the video?", + "choice_a": "seabed", + "choice_b": "in the air", + "choice_c": "aquatic", + "choice_d": "beach", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16315 + }, + { + "path": "39366.flac", + "question": "Where did the video take place?", + "choice_a": "indoor", + "choice_b": "forest", + "choice_c": "vegetable garden", + "choice_d": "The park is just around the corner", + "answer_gt": "forest", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16316 + }, + { + "path": "15226.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Television", + "choice_b": "speech", + "choice_c": "radio", + "choice_d": "Knock on the glass", + "answer_gt": "speech", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16317 + }, + { + "path": "9944.flac", + "question": "What are the people doing in the video?", + "choice_a": "Ride a motorcycle", + "choice_b": "racing", + "choice_c": "Drive a carriage", + "choice_d": "Driving a tractor", + "answer_gt": "Ride a motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16318 + }, + { + "path": "32879.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "dove", + "choice_c": "owl", + "choice_d": "wild goose", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16319 + }, + { + "path": "4841.flac", + "question": "What animal is in the video?", + "choice_a": "penguin", + "choice_b": "owl", + "choice_c": "rowboat", + "choice_d": "Petrel", + "answer_gt": "penguin", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16320 + }, + { + "path": "2428.flac", + "question": "What is the main sound source of the video?", + "choice_a": "honeybee", + "choice_b": "mouse", + "choice_c": "cricket", + "choice_d": "cicada", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16321 + }, + { + "path": "59540.flac", + "question": "What is the person in the video holding?", + "choice_a": "Electric shaver", + "choice_b": "Electric toothbrush", + "choice_c": "Electric hairdresser", + "choice_d": "Hair dryer", + "answer_gt": "Electric hairdresser", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16322 + }, + { + "path": "57990.flac", + "question": "What animal is in the video?", + "choice_a": "crow", + "choice_b": "chick", + "choice_c": "Oriole", + "choice_d": "Woodpecker", + "answer_gt": "Woodpecker", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16323 + }, + { + "path": "17827.flac", + "question": "Where does the video take place?", + "choice_a": "aquatic", + "choice_b": "in the water", + "choice_c": "at sea", + "choice_d": "highway", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16324 + }, + { + "path": "43334.flac", + "question": "Where did the video take place?", + "choice_a": "street", + "choice_b": "toilet", + "choice_c": "highway", + "choice_d": "Parking lot", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16325 + }, + { + "path": "6098.flac", + "question": "What are the people doing in the video?", + "choice_a": "Handmade", + "choice_b": "Drawing", + "choice_c": "write", + "choice_d": "Making toys", + "answer_gt": "Drawing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16326 + }, + { + "path": "22329.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16327 + }, + { + "path": "7165.flac", + "question": "What are the people doing in the video?", + "choice_a": "performing oral skills", + "choice_b": "drive a nail", + "choice_c": "Machine gun fire", + "choice_d": "sling", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16328 + }, + { + "path": "60625.flac", + "question": "What are the people doing in the video?", + "choice_a": "shuttlecock", + "choice_b": "play table tennis", + "choice_c": "Boxing", + "choice_d": "Play billiards", + "answer_gt": "play table tennis", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16329 + }, + { + "path": "3036.flac", + "question": "Where are the people in the video?", + "choice_a": "On the train", + "choice_b": "on board", + "choice_c": "On the plane", + "choice_d": "in the sea", + "answer_gt": "on board", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16330 + }, + { + "path": "43863.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "train", + "choice_b": "aircraft", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16331 + }, + { + "path": "18325.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16332 + }, + { + "path": "8680.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "police car siren", + "choice_b": "aircraft", + "choice_c": "ambulance", + "choice_d": "motorboat", + "answer_gt": "ambulance", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16333 + }, + { + "path": "24691.flac", + "question": "What animal is in the video?", + "choice_a": "chicken", + "choice_b": "bird", + "choice_c": "duck", + "choice_d": "cat", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16334 + }, + { + "path": "26641.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "donkey", + "choice_d": "rabbit", + "answer_gt": "donkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16335 + }, + { + "path": "43720.flac", + "question": "What's in the person's hand in the video?", + "choice_a": "Pen", + "choice_b": "Pistol", + "choice_c": "lighter", + "choice_d": "Water cup", + "answer_gt": "Pistol", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16336 + }, + { + "path": "9663.flac", + "question": "Where did the video take place?", + "choice_a": "On the road", + "choice_b": "seabed", + "choice_c": "of one's own unit", + "choice_d": "heavenly", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16337 + }, + { + "path": "8553.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "The cat barks", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16338 + }, + { + "path": "50335.flac", + "question": "What happened in the video?", + "choice_a": "pump", + "choice_b": "fire", + "choice_c": "volcanic explosion", + "choice_d": "storm", + "answer_gt": "fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16339 + }, + { + "path": "35102.flac", + "question": "What animal in the video makes a sound?", + "choice_a": "Cicada", + "choice_b": "cicada", + "choice_c": "grasshopper", + "choice_d": "cricket", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16340 + }, + { + "path": "46350.flac", + "question": "What are the people in the video doing?", + "choice_a": "drink water", + "choice_b": "Eat something", + "choice_c": "speech", + "choice_d": "chew", + "answer_gt": "Eat something", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16341 + }, + { + "path": "4836.flac", + "question": "What animal is in the video?", + "choice_a": "planing", + "choice_b": "honeybee", + "choice_c": "mosquito", + "choice_d": "fireworks and firecrackers", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16342 + }, + { + "path": "39722.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Car", + "choice_b": "gun", + "choice_c": "train", + "choice_d": "aircraft", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16343 + }, + { + "path": "47170.flac", + "question": "Where did the video take place?", + "choice_a": "square", + "choice_b": "In the forest", + "choice_c": "toilet", + "choice_d": "grassland", + "answer_gt": "In the forest", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16344 + }, + { + "path": "10100.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "police car siren", + "choice_b": "owl", + "choice_c": "chick", + "choice_d": "Birdsong", + "answer_gt": "Birdsong", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16345 + }, + { + "path": "43885.flac", + "question": "What's the weather in the video?", + "choice_a": "overcast", + "choice_b": "a sunny day", + "choice_c": "cloudy", + "choice_d": "rain", + "answer_gt": "a sunny day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16346 + }, + { + "path": "45604.flac", + "question": "What animal appears in the video?", + "choice_a": "dove", + "choice_b": "honeybee", + "choice_c": "parrot", + "choice_d": "elephant", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16347 + }, + { + "path": "59003.flac", + "question": "What is shown in the video?", + "choice_a": "Toilet flush", + "choice_b": "The toilet is clogged.", + "choice_c": "Sewer flushing", + "choice_d": "The wash basin is leaking", + "answer_gt": "Toilet flush", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16348 + }, + { + "path": "37977.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "pig", + "choice_c": "cat", + "choice_d": "chipmunks", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16349 + }, + { + "path": "13998.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "bird", + "choice_d": "fox", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16350 + }, + { + "path": "43538.flac", + "question": "What is he doing in the video?", + "choice_a": "Cutting cloth", + "choice_b": "Cut the bottle", + "choice_c": "Making toys", + "choice_d": "Tearing paper", + "answer_gt": "Cut the bottle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16351 + }, + { + "path": "29330.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "motorboat", + "choice_d": "Bus", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16352 + }, + { + "path": "5638.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "racing", + "choice_d": "bus", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16353 + }, + { + "path": "45448.flac", + "question": "Where did the video take place?", + "choice_a": "Swimming Pool", + "choice_b": "forest", + "choice_c": "surface of the water", + "choice_d": "Underwater", + "answer_gt": "Underwater", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16354 + }, + { + "path": "24682.flac", + "question": "What is the main sound source of the video?", + "choice_a": "dog", + "choice_b": "frog", + "choice_c": "crocodile", + "choice_d": "Toad", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16355 + }, + { + "path": "57734.flac", + "question": "What are the people doing in the video?", + "choice_a": "Tractor ploughing", + "choice_b": "Driving the harvester", + "choice_c": "Earth digging by excavator", + "choice_d": "Driving a lawn mower", + "answer_gt": "Driving a lawn mower", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16356 + }, + { + "path": "15082.flac", + "question": "What are the people doing in the video?", + "choice_a": "Machine gun fire", + "choice_b": "performing oral skills", + "choice_c": "go hunting", + "choice_d": "Shooting", + "answer_gt": "go hunting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16357 + }, + { + "path": "30733.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16358 + }, + { + "path": "60870.flac", + "question": "What are the people doing in the video?", + "choice_a": "haircut", + "choice_b": "to shave", + "choice_c": "wash one 's face", + "choice_d": "Nose hair repair", + "answer_gt": "to shave", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16359 + }, + { + "path": "52948.flac", + "question": "Where did the video take place?", + "choice_a": "Swimming Pool", + "choice_b": "kitchen", + "choice_c": "Bathroom", + "choice_d": "TOILET", + "answer_gt": "TOILET", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16360 + }, + { + "path": "59615.flac", + "question": "Where does the animal in the video take place?", + "choice_a": "chicken farm", + "choice_b": "Duck farm", + "choice_c": "Goose farm", + "choice_d": "Turkey farm", + "answer_gt": "chicken farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16361 + }, + { + "path": "14140.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Fire engine siren", + "choice_b": "bus", + "choice_c": "Ambulance whistle", + "choice_d": "Civil defense alarm", + "answer_gt": "Ambulance whistle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16362 + }, + { + "path": "36490.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "rabbit", + "choice_c": "cat", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16363 + }, + { + "path": "56344.flac", + "question": "What are the people doing in the video?", + "choice_a": "Practice playing volleyball", + "choice_b": "Practice playing badminton", + "choice_c": "Practice Bowling", + "choice_d": "sprint", + "answer_gt": "Practice playing badminton", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16364 + }, + { + "path": "41151.flac", + "question": "What animal appears in the video?", + "choice_a": "tiger", + "choice_b": "lion", + "choice_c": "leopard", + "choice_d": "Cattle", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16365 + }, + { + "path": "13413.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "chicken", + "choice_c": "cat", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16366 + }, + { + "path": "30644.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "Hippo", + "choice_c": "rhinoceros", + "choice_d": "elephant", + "answer_gt": "elephant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16367 + }, + { + "path": "54132.flac", + "question": "What is the main sound source in this video?", + "choice_a": "lawn mower", + "choice_b": "The sound of vacuum cleaner cleaning the floor", + "choice_c": "sweeping machine", + "choice_d": "The sound of hair dryer", + "answer_gt": "The sound of vacuum cleaner cleaning the floor", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16368 + }, + { + "path": "35510.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Blow the trumpet", + "choice_b": "singing", + "choice_c": "Playing flute", + "choice_d": "Car horn", + "answer_gt": "Car horn", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16369 + }, + { + "path": "38163.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of rain", + "choice_b": "The sound of a waterfall", + "choice_c": "shot", + "choice_d": "thunder", + "answer_gt": "thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16370 + }, + { + "path": "51092.flac", + "question": "What is the occupation of the person in the video?", + "choice_a": "migrant workers", + "choice_b": "Sanitation worker", + "choice_c": "cleaner", + "choice_d": "Property workers", + "answer_gt": "cleaner", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16371 + }, + { + "path": "4522.flac", + "question": "What's flying in the sky?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "aircraft", + "choice_d": "parachute", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16372 + }, + { + "path": "27061.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "motorcycle", + "choice_b": "helicopter", + "choice_c": "motorboat", + "choice_d": "bus", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16373 + }, + { + "path": "32872.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "turkey", + "choice_c": "duck", + "choice_d": "dove", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16374 + }, + { + "path": "41248.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "the sound of driving a snowmobile", + "choice_d": "lawn mower", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16375 + }, + { + "path": "43764.flac", + "question": "Where are the animals in the video?", + "choice_a": "On the branch", + "choice_b": "In the cage", + "choice_c": "On the pole", + "choice_d": "In the henhouse", + "answer_gt": "On the branch", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16376 + }, + { + "path": "11550.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Smoke Alarm", + "choice_b": "The sound of the trumpet", + "choice_c": "Fire engine whistle", + "choice_d": "Civil defense alarm", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16377 + }, + { + "path": "24520.flac", + "question": "What animal appears in the video?", + "choice_a": "penguin", + "choice_b": "dog", + "choice_c": "swan", + "choice_d": "cockcrow", + "answer_gt": "penguin", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16378 + }, + { + "path": "48123.flac", + "question": "What happened to the video?", + "choice_a": "pump", + "choice_b": "Volcanic explosion", + "choice_c": "tsunami", + "choice_d": "sand storm", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16379 + }, + { + "path": "32986.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of rain", + "choice_b": "aircraft", + "choice_c": "Underwater sound", + "choice_d": "thunder", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16380 + }, + { + "path": "37665.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "bus", + "choice_d": "motorcycle", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16381 + }, + { + "path": "10980.flac", + "question": "What's driving in the video?", + "choice_a": "lathe rotation sound", + "choice_b": "motorcycle", + "choice_c": "highway", + "choice_d": "automobile", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16382 + }, + { + "path": "33200.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "automobile", + "choice_c": "glider", + "choice_d": "Bus", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16383 + }, + { + "path": "22172.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "helicopter", + "choice_c": "acceleration sound of automobile engine", + "choice_d": "gun", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16384 + }, + { + "path": "19781.flac", + "question": "What is the main sound source of the video?", + "choice_a": "highway", + "choice_b": "aircraft", + "choice_c": "bus", + "choice_d": "rocket launching", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16385 + }, + { + "path": "38969.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "chipmunks", + "choice_c": "fox", + "choice_d": "cat", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16386 + }, + { + "path": "48247.flac", + "question": "Where did the video take place?", + "choice_a": "seabed", + "choice_b": "aquatic", + "choice_c": "toilet", + "choice_d": "natatorium", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16387 + }, + { + "path": "8897.flac", + "question": "What animal appears in the video?", + "choice_a": "parrot", + "choice_b": "Thrush", + "choice_c": "crow", + "choice_d": "tiger", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16388 + }, + { + "path": "55819.flac", + "question": "What are the people doing in the video?", + "choice_a": "play volleyball", + "choice_b": "play tennis", + "choice_c": "Play basketball", + "choice_d": "Play billiards", + "answer_gt": "play volleyball", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16389 + }, + { + "path": "31682.flac", + "question": "What animal appears in the video?", + "choice_a": "Myna", + "choice_b": "chicken", + "choice_c": "honeybee", + "choice_d": "Canary", + "answer_gt": "Myna", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16390 + }, + { + "path": "29919.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "chicken", + "choice_c": "dove", + "choice_d": "duck", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16391 + }, + { + "path": "21955.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "duck", + "choice_c": "bird", + "choice_d": "dog", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16392 + }, + { + "path": "17798.flac", + "question": "What is the main sound source of the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "cattle", + "choice_d": "dog", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16393 + }, + { + "path": "1664.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "automobile", + "choice_b": "motorcycle", + "choice_c": "train", + "choice_d": "helicopter", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16394 + }, + { + "path": "17896.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "performing oral skills", + "choice_b": "Block shot", + "choice_c": "Shooting", + "choice_d": "To strike a fire", + "answer_gt": "Block shot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16395 + }, + { + "path": "33610.flac", + "question": "What are the people doing in the video?", + "choice_a": "woodpeckers peck at trees", + "choice_b": "Using the loom", + "choice_c": "Using a sewing machine", + "choice_d": "Using cutting machine", + "answer_gt": "Using a sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16396 + }, + { + "path": "16345.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "fox", + "choice_c": "dog", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16397 + }, + { + "path": "12993.flac", + "question": "What are the people in the video doing?", + "choice_a": "watering", + "choice_b": "performing oral skills", + "choice_c": "Gun fire", + "choice_d": "Shooting", + "answer_gt": "Gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16398 + }, + { + "path": "46854.flac", + "question": "What are the people in the video doing?", + "choice_a": "plough fields", + "choice_b": "Sweep the floor", + "choice_c": "want to buy", + "choice_d": "cut grass", + "answer_gt": "cut grass", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16399 + }, + { + "path": "60223.flac", + "question": "What happened in the video?", + "choice_a": "New year's Eve", + "choice_b": "Setting off fireworks", + "choice_c": "June 1 Carnival", + "choice_d": "The comet passed by", + "answer_gt": "Setting off fireworks", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16400 + }, + { + "path": "51701.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "Tractor", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16401 + }, + { + "path": "2074.flac", + "question": "What happened in the video?", + "choice_a": "Feeding", + "choice_b": "Dogs eat food", + "choice_c": "keep a dog", + "choice_d": "The dog gasped", + "answer_gt": "Feeding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16402 + }, + { + "path": "59632.flac", + "question": "What are the people doing in the video?", + "choice_a": "Teaching and playing basketball", + "choice_b": "Teaching baseball", + "choice_c": "Teaching table tennis", + "choice_d": "Teaching and playing volleyball", + "answer_gt": "Teaching and playing volleyball", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16403 + }, + { + "path": "23061.flac", + "question": "What are the people in the video doing?", + "choice_a": "Play with fire", + "choice_b": "smoking", + "choice_c": "Light firecrackers", + "choice_d": "sparking", + "answer_gt": "Light firecrackers", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16404 + }, + { + "path": "5566.flac", + "question": "What is the main sound source of the video?", + "choice_a": "motorcycle", + "choice_b": "train", + "choice_c": "automobile", + "choice_d": "ship", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16405 + }, + { + "path": "3591.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "crow", + "choice_c": "dove", + "choice_d": "owl", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16406 + }, + { + "path": "57979.flac", + "question": "What is shown in the video?", + "choice_a": "scenery", + "choice_b": "Vehicle technology", + "choice_c": "to catch fire", + "choice_d": "light stick", + "answer_gt": "Vehicle technology", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16407 + }, + { + "path": "46359.flac", + "question": "Where did the video take place?", + "choice_a": "In the water", + "choice_b": "grassland", + "choice_c": "In the park", + "choice_d": "By the beach", + "answer_gt": "In the water", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16408 + }, + { + "path": "24923.flac", + "question": "What is the main sound source of the video?", + "choice_a": "automobile", + "choice_b": "aircraft", + "choice_c": "thunder", + "choice_d": "bus", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16409 + }, + { + "path": "37838.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "helicopter", + "choice_b": "train", + "choice_c": "the sound of driving a snowmobile", + "choice_d": "gun", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16410 + }, + { + "path": "57617.flac", + "question": "Where did the video take place?", + "choice_a": "workshop", + "choice_b": "factory", + "choice_c": "repair depot", + "choice_d": "Warehouse", + "answer_gt": "workshop", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16411 + }, + { + "path": "15833.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "cricket", + "choice_b": "dog", + "choice_c": "fireworks and firecrackers", + "choice_d": "mantis", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16412 + }, + { + "path": "6666.flac", + "question": "What are the people doing in this video?", + "choice_a": "sailing", + "choice_b": "diving", + "choice_c": "setting off fireworks", + "choice_d": "Play on the water slide", + "answer_gt": "Play on the water slide", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16413 + }, + { + "path": "28422.flac", + "question": "What is the main sound source of the video?", + "choice_a": "car engine noise", + "choice_b": "thunder", + "choice_c": "lawn mower", + "choice_d": "hail", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16414 + }, + { + "path": "49025.flac", + "question": "What are the people in the video doing?", + "choice_a": "fly an airplane", + "choice_b": "Driving a racing car", + "choice_c": "Driving a tractor", + "choice_d": "Driving a train", + "answer_gt": "Driving a racing car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16415 + }, + { + "path": "26822.flac", + "question": "Where did the video take place?", + "choice_a": "In the forest", + "choice_b": "playground", + "choice_c": "desert", + "choice_d": "sandy beach", + "answer_gt": "sandy beach", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16416 + }, + { + "path": "59753.flac", + "question": "What animal is in the video?", + "choice_a": "parrot", + "choice_b": "Woodpecker", + "choice_c": "Oriole", + "choice_d": "canary", + "answer_gt": "Woodpecker", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16417 + }, + { + "path": "1743.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16418 + }, + { + "path": "45321.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The cow barked", + "choice_b": "dog's bark", + "choice_c": "police car siren", + "choice_d": "the lion roared", + "answer_gt": "dog's bark", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16419 + }, + { + "path": "10701.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "honeybee", + "choice_b": "cricket", + "choice_c": "fireworks and firecrackers", + "choice_d": "Cicadas call", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16420 + }, + { + "path": "48416.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "tram", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16421 + }, + { + "path": "7928.flac", + "question": "What are the people doing in the video?", + "choice_a": "racing", + "choice_b": "typing on a typewriter", + "choice_c": "play with water", + "choice_d": "skiing", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16422 + }, + { + "path": "14861.flac", + "question": "What are the people doing in the video?", + "choice_a": "sailing", + "choice_b": "Hold your breath", + "choice_c": "diving", + "choice_d": "Catching fish", + "answer_gt": "diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16423 + }, + { + "path": "28959.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "dog", + "choice_d": "pig", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16424 + }, + { + "path": "23352.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of wind", + "choice_b": "hail", + "choice_c": "acceleration sound of automobile engine", + "choice_d": "sound of rain", + "answer_gt": "sound of rain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16425 + }, + { + "path": "42820.flac", + "question": "Where does the video take place?", + "choice_a": "street", + "choice_b": "toilet", + "choice_c": "forest", + "choice_d": "grassland", + "answer_gt": "grassland", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16426 + }, + { + "path": "48259.flac", + "question": "What animal appears in the video?", + "choice_a": "lion", + "choice_b": "elephant", + "choice_c": "leopard", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16427 + }, + { + "path": "6809.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "motorboat", + "choice_c": "Driving a motorcycle", + "choice_d": "Harvester", + "answer_gt": "Driving a motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16428 + }, + { + "path": "9912.flac", + "question": "What animal appears in the video?", + "choice_a": "dove", + "choice_b": "owl", + "choice_c": "ostrich", + "choice_d": "pheasant", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16429 + }, + { + "path": "33893.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16430 + }, + { + "path": "21632.flac", + "question": "What are the people in the video doing?", + "choice_a": "Parachute jump", + "choice_b": "fly a kite", + "choice_c": "skiing", + "choice_d": "remote controlled aircraft", + "answer_gt": "fly a kite", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16431 + }, + { + "path": "45203.flac", + "question": "What are the people in the video doing?", + "choice_a": "roller coaster running", + "choice_b": "Roller-skating", + "choice_c": "skating", + "choice_d": "skiing", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16432 + }, + { + "path": "2625.flac", + "question": "What is the purpose of the sound in the video?", + "choice_a": "decompression", + "choice_b": "interesting", + "choice_c": "entertainment", + "choice_d": "Program effect", + "answer_gt": "decompression", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16433 + }, + { + "path": "19515.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "dove", + "choice_c": "dog", + "choice_d": "peacock", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16434 + }, + { + "path": "24504.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "bird", + "choice_d": "Hyenas", + "answer_gt": "Hyenas", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16435 + }, + { + "path": "33555.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "chipmunks", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16436 + }, + { + "path": "26311.flac", + "question": "What are the people in the video doing?", + "choice_a": "Swimming", + "choice_b": "diving", + "choice_c": "go fishing", + "choice_d": "dance", + "answer_gt": "diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16437 + }, + { + "path": "1234.flac", + "question": "What happened in the video?", + "choice_a": "Volcano eruption", + "choice_b": "tornado", + "choice_c": "blast", + "choice_d": "cars crush things", + "answer_gt": "blast", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16438 + }, + { + "path": "30798.flac", + "question": "Where did the video take place?", + "choice_a": "mountain path", + "choice_b": "playground", + "choice_c": "On the road", + "choice_d": "forest", + "answer_gt": "On the road", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16439 + }, + { + "path": "50599.flac", + "question": "Where did the video take place?", + "choice_a": "On the grass", + "choice_b": "On the road", + "choice_c": "outskirts", + "choice_d": "On the ranch", + "answer_gt": "On the grass", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16440 + }, + { + "path": "52972.flac", + "question": "What are the people in the video doing?", + "choice_a": "Driving a motorcycle", + "choice_b": "Driving a racing car", + "choice_c": "Driving a bus", + "choice_d": "Yachting", + "answer_gt": "Driving a racing car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16441 + }, + { + "path": "17884.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "ship", + "choice_d": "bus", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16442 + }, + { + "path": "60724.flac", + "question": "What are the people doing in the video?", + "choice_a": "And cement", + "choice_b": "Porcelain drill", + "choice_c": "House decoration", + "choice_d": "Vacuum the floor", + "answer_gt": "Vacuum the floor", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16443 + }, + { + "path": "14645.flac", + "question": "What are the people doing in the video?", + "choice_a": "sailing", + "choice_b": "diving", + "choice_c": "flush the toilet", + "choice_d": "Hold your breath", + "answer_gt": "diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16444 + }, + { + "path": "25906.flac", + "question": "What's the weather in the video?", + "choice_a": "overcast", + "choice_b": "A snowy day", + "choice_c": "rain", + "choice_d": "a sunny day", + "answer_gt": "a sunny day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16445 + }, + { + "path": "18116.flac", + "question": "What are the people in the video doing?", + "choice_a": "roller coaster running", + "choice_b": "bullfight", + "choice_c": "Riding cattle", + "choice_d": "riding", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16446 + }, + { + "path": "26521.flac", + "question": "Where did the video take place?", + "choice_a": "On the road", + "choice_b": "playground", + "choice_c": "seaside", + "choice_d": "ski field", + "answer_gt": "seaside", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16447 + }, + { + "path": "2280.flac", + "question": "What is the main sound source of the video?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "aircraft", + "choice_d": "motorcycle", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16448 + }, + { + "path": "33156.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "Sewing machine", + "choice_c": "stapler", + "choice_d": "Dust collector", + "answer_gt": "Sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16449 + }, + { + "path": "30391.flac", + "question": "What animal appears in the video?", + "choice_a": "dove", + "choice_b": "peacock", + "choice_c": "pheasant", + "choice_d": "ostrich", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16450 + }, + { + "path": "28509.flac", + "question": "What is the main sound source of the video?", + "choice_a": "lathe rotation sound", + "choice_b": "aircraft", + "choice_c": "bus", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16451 + }, + { + "path": "27868.flac", + "question": "What animal appears in the video?", + "choice_a": "bird", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "dog", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16452 + }, + { + "path": "51575.flac", + "question": "What happened in the video?", + "choice_a": "duck", + "choice_b": "fire", + "choice_c": "Burning fireworks", + "choice_d": "Liquid explosion", + "answer_gt": "fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16453 + }, + { + "path": "15109.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "goose", + "choice_b": "chicken", + "choice_c": "swallow", + "choice_d": "duck", + "answer_gt": "goose", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16454 + }, + { + "path": "29941.flac", + "question": "What is the main sound source of the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "cattle", + "choice_d": "sheep", + "answer_gt": "donkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16455 + }, + { + "path": "53118.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "one-arm bandit", + "choice_b": "Television", + "choice_c": "recreational machines", + "choice_d": "Basketball machine", + "answer_gt": "one-arm bandit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16456 + }, + { + "path": "5455.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "motorboat", + "choice_c": "seeder", + "choice_d": "helicopter", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16457 + }, + { + "path": "61546.flac", + "question": "What are the people doing in the video?", + "choice_a": "Latin dance", + "choice_b": "Hip hop", + "choice_c": "Classical dance", + "choice_d": "Belly dancing", + "answer_gt": "Hip hop", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16458 + }, + { + "path": "2398.flac", + "question": "When is it in the video?", + "choice_a": "day", + "choice_b": "morning", + "choice_c": "dawn", + "choice_d": "evening", + "answer_gt": "evening", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16459 + }, + { + "path": "59426.flac", + "question": "What is shown in the video?", + "choice_a": "Electric toothbrush", + "choice_b": "Electric hairdresser", + "choice_c": "Electric shaver", + "choice_d": "Hair dryer", + "answer_gt": "Electric shaver", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16460 + }, + { + "path": "18598.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "waterfall", + "choice_b": "sound of wind", + "choice_c": "car engine noise", + "choice_d": "sea wave", + "answer_gt": "sea wave", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16461 + }, + { + "path": "43963.flac", + "question": "What's the weather in the video?", + "choice_a": "overcast", + "choice_b": "a sunny day", + "choice_c": "rain", + "choice_d": "snowy day", + "answer_gt": "a sunny day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16462 + }, + { + "path": "8400.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "a car", + "choice_b": "The sound of running water", + "choice_c": "Subway horn", + "choice_d": "Radio-controlled helicopte", + "answer_gt": "Radio-controlled helicopte", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16463 + }, + { + "path": "17439.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "racing", + "choice_b": "motorcycle", + "choice_c": "fireworks and firecrackers", + "choice_d": "Tricycle", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16464 + }, + { + "path": "59831.flac", + "question": "What happened in the video?", + "choice_a": "Bulldozer", + "choice_b": "Excavators break stones", + "choice_c": "Freight transport", + "choice_d": "Crane lifting things", + "answer_gt": "Excavators break stones", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16465 + }, + { + "path": "6881.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "cattle", + "choice_d": "camel", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16466 + }, + { + "path": "44867.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Fire engine siren", + "choice_b": "honk", + "choice_c": "vocal imitation performance", + "choice_d": "Alarm", + "answer_gt": "Alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16467 + }, + { + "path": "19342.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "car engine noise", + "choice_c": "The sound of waves", + "choice_d": "Airplane noise", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16468 + }, + { + "path": "54728.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The tiger barks", + "choice_b": "The lion barked", + "choice_c": "The bear barked", + "choice_d": "Cheetah calls", + "answer_gt": "Cheetah calls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16469 + }, + { + "path": "9308.flac", + "question": "What are the people doing in the video?", + "choice_a": "racing", + "choice_b": "skiing", + "choice_c": "skateboarding", + "choice_d": "Chasing chicken", + "answer_gt": "Chasing chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16470 + }, + { + "path": "53337.flac", + "question": "What are the people in the video doing?", + "choice_a": "Cutting cloth", + "choice_b": "paper-cut", + "choice_c": "Tearing paper", + "choice_d": "Origami Aircraft", + "answer_gt": "Tearing paper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16471 + }, + { + "path": "57410.flac", + "question": "Where is the bird in the video?", + "choice_a": "Bird's nest", + "choice_b": "In the birdcage", + "choice_c": "In the wild", + "choice_d": "On the grass", + "answer_gt": "In the birdcage", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16472 + }, + { + "path": "42466.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "thunder", + "choice_b": "firecrackers", + "choice_c": "The sound of the trumpet", + "choice_d": "shot", + "answer_gt": "firecrackers", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16473 + }, + { + "path": "50899.flac", + "question": "Where is the bird in the video?", + "choice_a": "In a birdcage", + "choice_b": "In the grass", + "choice_c": "On the branch", + "choice_d": "On the leaves", + "answer_gt": "On the branch", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16474 + }, + { + "path": "56315.flac", + "question": "Where did the video take place?", + "choice_a": "classroom", + "choice_b": "Barber Shop", + "choice_c": "cinema", + "choice_d": "restaurant", + "answer_gt": "Barber Shop", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16475 + }, + { + "path": "7665.flac", + "question": "What animals appear in the video?", + "choice_a": "chicken", + "choice_b": "dog", + "choice_c": "owl", + "choice_d": "peacock", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16476 + }, + { + "path": "4370.flac", + "question": "What are the people doing in this video?", + "choice_a": "Nailing", + "choice_b": "performing oral skills", + "choice_c": "rope skipping", + "choice_d": "Shooting", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16477 + }, + { + "path": "45981.flac", + "question": "What happened in the video?", + "choice_a": "Hunting", + "choice_b": "Dogs and magpies make a scene", + "choice_c": "Cat and duck fight", + "choice_d": "A fight between a chicken and a sheep", + "answer_gt": "Dogs and magpies make a scene", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16478 + }, + { + "path": "27764.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "bird", + "choice_c": "duck", + "choice_d": "dog", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16479 + }, + { + "path": "31447.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "dog", + "choice_c": "cat", + "choice_d": "rabbit", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16480 + }, + { + "path": "10924.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "highway", + "choice_c": "motorcycle", + "choice_d": "motorboat", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16481 + }, + { + "path": "43082.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "train", + "choice_b": "sliding door", + "choice_c": "aircraft", + "choice_d": "gun", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16482 + }, + { + "path": "50376.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "Car", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16483 + }, + { + "path": "46701.flac", + "question": "What animal appears in the video?", + "choice_a": "planing", + "choice_b": "elephant", + "choice_c": "mantis", + "choice_d": "honeybee", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16484 + }, + { + "path": "6363.flac", + "question": "What animal is in the video?", + "choice_a": "dove", + "choice_b": "owl", + "choice_c": "chick", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16485 + }, + { + "path": "11989.flac", + "question": "What animal is in the video?", + "choice_a": "frog", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "dove", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16486 + }, + { + "path": "26630.flac", + "question": "What are the people in the video doing?", + "choice_a": "Diving", + "choice_b": "Swimming", + "choice_c": "dance", + "choice_d": "High jump", + "answer_gt": "Diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16487 + }, + { + "path": "46546.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "rope skipping", + "choice_b": "thunder", + "choice_c": "hail", + "choice_d": "The sound of rain\u00b7", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16488 + }, + { + "path": "45018.flac", + "question": "What animal appears in the video?", + "choice_a": "Myna", + "choice_b": "dove", + "choice_c": "elephant", + "choice_d": "parrot", + "answer_gt": "Myna", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16489 + }, + { + "path": "21801.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "wolf", + "choice_c": "tiger", + "choice_d": "lion", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16490 + }, + { + "path": "22394.flac", + "question": "Where does the video take place?", + "choice_a": "aquatic", + "choice_b": "at sea", + "choice_c": "field", + "choice_d": "woods", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16491 + }, + { + "path": "44527.flac", + "question": "Where did the video take place?", + "choice_a": "zoo", + "choice_b": "park", + "choice_c": "Football Field", + "choice_d": "Racetrack", + "answer_gt": "Racetrack", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16492 + }, + { + "path": "23481.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "wolf", + "choice_d": "elephant", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16493 + }, + { + "path": "46405.flac", + "question": "Where are the cows in the video?", + "choice_a": "grassland", + "choice_b": "street", + "choice_c": "toilet", + "choice_d": "sandy beach", + "answer_gt": "grassland", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16494 + }, + { + "path": "24671.flac", + "question": "What are the people in the video doing?", + "choice_a": "skiing", + "choice_b": "bullfight", + "choice_c": "Riding a donkey", + "choice_d": "riding", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16495 + }, + { + "path": "51298.flac", + "question": "What is the number of marks on the plane in the video?", + "choice_a": "0475", + "choice_b": "0457", + "choice_c": "one thousand two hundred and thirty-four", + "choice_d": "two thousand five hundred and sixty-seven", + "answer_gt": "0475", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16496 + }, + { + "path": "45178.flac", + "question": "Where did the video take place?", + "choice_a": "street", + "choice_b": "grassland", + "choice_c": "desert", + "choice_d": "Grassland", + "answer_gt": "grassland", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16497 + }, + { + "path": "44638.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Sheep crow", + "choice_b": "The cat barks", + "choice_c": "dog's bark", + "choice_d": "The sound of a fox", + "answer_gt": "dog's bark", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16498 + }, + { + "path": "14578.flac", + "question": "What are the people doing in the video?", + "choice_a": "skiing", + "choice_b": "typing on a typewriter", + "choice_c": "play with water", + "choice_d": "get on", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16499 + }, + { + "path": "51954.flac", + "question": "How do the animals in the video keep a movement?", + "choice_a": "Get hurt", + "choice_b": "lazy", + "choice_c": "Watching the birds in the sky", + "choice_d": "fell asleep", + "answer_gt": "fell asleep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16500 + }, + { + "path": "35385.flac", + "question": "What happened in the video?", + "choice_a": "A fight between a dog and a cat", + "choice_b": "The dog found the snake", + "choice_c": "The dog was bitten by a snake", + "choice_d": "The dog quarreled", + "answer_gt": "The dog found the snake", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16501 + }, + { + "path": "58145.flac", + "question": "What happened in the video?", + "choice_a": "high jump", + "choice_b": "Performing acrobatics", + "choice_c": "Skateboard wrestling", + "choice_d": "Play balance car wrestling", + "answer_gt": "Skateboard wrestling", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16502 + }, + { + "path": "16597.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "Car driving", + "choice_c": "Underwater sound", + "choice_d": "vacuum cleaner to clean the floor", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16503 + }, + { + "path": "50874.flac", + "question": "Where did the video take place?", + "choice_a": "On the ship", + "choice_b": "On the bus", + "choice_c": "In the hotel", + "choice_d": "On the plane", + "answer_gt": "On the plane", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16504 + }, + { + "path": "25198.flac", + "question": "What happened in the video?", + "choice_a": "Set off firecrackers", + "choice_b": "The car is on fire", + "choice_c": "fire", + "choice_d": "The train exploded", + "answer_gt": "fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16505 + }, + { + "path": "49259.flac", + "question": "What are the people in the video doing?", + "choice_a": "plough fields", + "choice_b": "Cultivated land", + "choice_c": "cut grass", + "choice_d": "want to buy", + "answer_gt": "cut grass", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16506 + }, + { + "path": "37352.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "sheep", + "choice_d": "dog", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16507 + }, + { + "path": "53356.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of wind", + "choice_b": "The fox barked", + "choice_c": "The sound of sheep biting", + "choice_d": "The sound of dog biting", + "answer_gt": "The sound of dog biting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16508 + }, + { + "path": "53871.flac", + "question": "Where does the video take place?", + "choice_a": "train station", + "choice_b": "Basketball Court", + "choice_c": "Racetrack", + "choice_d": "Football Field", + "answer_gt": "Racetrack", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16509 + }, + { + "path": "60403.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Play the recording", + "choice_b": "The sound of shacks", + "choice_c": "Blowing the trumpet", + "choice_d": "play vertical flute", + "answer_gt": "Blowing the trumpet", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16510 + }, + { + "path": "44283.flac", + "question": "Where does the video take place?", + "choice_a": "seabed", + "choice_b": "bathtub", + "choice_c": "on the subway", + "choice_d": "Swimming Pool", + "answer_gt": "Swimming Pool", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16511 + }, + { + "path": "11954.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Gun fire", + "choice_b": "firecrackers", + "choice_c": "blast", + "choice_d": "eat apples", + "answer_gt": "firecrackers", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16512 + }, + { + "path": "47297.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "machine", + "choice_b": "The sound of the trumpet", + "choice_c": "Machine sound", + "choice_d": "blender", + "answer_gt": "machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16513 + }, + { + "path": "59143.flac", + "question": "What are the people doing in the video?", + "choice_a": "applause", + "choice_b": "Mower mowing", + "choice_c": "cheer", + "choice_d": "revolving door", + "answer_gt": "cheer", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16514 + }, + { + "path": "58550.flac", + "question": "What are the people doing in the video?", + "choice_a": "play tennis", + "choice_b": "play table tennis", + "choice_c": "Play baseball", + "choice_d": "shuttlecock", + "answer_gt": "shuttlecock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16515 + }, + { + "path": "32768.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "waterfall", + "choice_b": "diving", + "choice_c": "thunder", + "choice_d": "sound of wind", + "answer_gt": "sound of wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16516 + }, + { + "path": "24440.flac", + "question": "What animal is in the video?", + "choice_a": "duck", + "choice_b": "goose", + "choice_c": "dove", + "choice_d": "penguin", + "answer_gt": "duck", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16517 + }, + { + "path": "26101.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "tiger", + "choice_c": "dog", + "choice_d": "The bear", + "answer_gt": "tiger", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16518 + }, + { + "path": "37813.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "train", + "choice_c": "the sound of driving a snowmobile", + "choice_d": "gun", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16519 + }, + { + "path": "18438.flac", + "question": "Where does the video take place?", + "choice_a": "Underwater", + "choice_b": "aquatic", + "choice_c": "ski field", + "choice_d": "at sea", + "answer_gt": "Underwater", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16520 + }, + { + "path": "12542.flac", + "question": "What are the people in the video doing?", + "choice_a": "racing", + "choice_b": "skiing", + "choice_c": "riding", + "choice_d": "bullfight", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16521 + }, + { + "path": "36062.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "parrot", + "choice_c": "Thrush", + "choice_d": "wild goose", + "answer_gt": "Thrush", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16522 + }, + { + "path": "12255.flac", + "question": "What animal appears in the video?", + "choice_a": "rowboat", + "choice_b": "a hippocampus", + "choice_c": "walrus", + "choice_d": "Sea lions", + "answer_gt": "Sea lions", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16523 + }, + { + "path": "45397.flac", + "question": "What animal appears in the video?", + "choice_a": "honeybee", + "choice_b": "crow", + "choice_c": "parrot", + "choice_d": "tiger", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16524 + }, + { + "path": "20473.flac", + "question": "Where does the video take place?", + "choice_a": "Swimming Pool", + "choice_b": "aquatic", + "choice_c": "at sea", + "choice_d": "of one's own unit", + "answer_gt": "Swimming Pool", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16525 + }, + { + "path": "35688.flac", + "question": "What animal appears in the video?", + "choice_a": "Otter", + "choice_b": "cricket", + "choice_c": "rowboat", + "choice_d": "Sea lions", + "answer_gt": "Otter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16526 + }, + { + "path": "3307.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "motorcycle", + "choice_c": "automobile", + "choice_d": "motorboat", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16527 + }, + { + "path": "35760.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "dog", + "choice_c": "wolf", + "choice_d": "pig", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16528 + }, + { + "path": "32881.flac", + "question": "What happened in the video?", + "choice_a": "Volcano eruption", + "choice_b": "Here comes the tornado", + "choice_c": "pump", + "choice_d": "Here comes the sandstorm", + "answer_gt": "Here comes the tornado", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16529 + }, + { + "path": "57616.flac", + "question": "What is shown in the video?", + "choice_a": "Floor sweeping robot", + "choice_b": "Toy train", + "choice_c": "Vacuum cleaner", + "choice_d": "tank", + "answer_gt": "Floor sweeping robot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16530 + }, + { + "path": "45493.flac", + "question": "What animal appears in the video?", + "choice_a": "elephant", + "choice_b": "cattle", + "choice_c": "The bear", + "choice_d": "lion", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16531 + }, + { + "path": "56894.flac", + "question": "What happened in the video?", + "choice_a": "Oriole Birds eat", + "choice_b": "repeat the words of others like a parrot", + "choice_c": "Woodpeckers peck at trees", + "choice_d": "The self-service feeder works normally", + "answer_gt": "Oriole Birds eat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16532 + }, + { + "path": "11669.flac", + "question": "What animals appear in the video?", + "choice_a": "dog", + "choice_b": "sheep", + "choice_c": "duck", + "choice_d": "cat", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16533 + }, + { + "path": "8940.flac", + "question": "What is the main sound source of the video?", + "choice_a": "vacuum cleaner to clean the floor", + "choice_b": "Music box", + "choice_c": "Harmonica sound", + "choice_d": "Wind chime", + "answer_gt": "Wind chime", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16534 + }, + { + "path": "20472.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Wind Chime", + "choice_b": "train", + "choice_c": "metro", + "choice_d": "bus", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16535 + }, + { + "path": "49310.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Train horn", + "choice_b": "Airplane noise", + "choice_c": "Bus horn", + "choice_d": "Ambulance siren", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16536 + }, + { + "path": "4280.flac", + "question": "What are the people doing in this video?", + "choice_a": "eat something", + "choice_b": "machine", + "choice_c": "skiing", + "choice_d": "Live broadcast", + "answer_gt": "Live broadcast", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16537 + }, + { + "path": "24944.flac", + "question": "What animal appears in the video?", + "choice_a": "goose", + "choice_b": "dog", + "choice_c": "chicken", + "choice_d": "eagle", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16538 + }, + { + "path": "8392.flac", + "question": "What are the people doing in the video?", + "choice_a": "riding", + "choice_b": "roller coaster running", + "choice_c": "race", + "choice_d": "Hunting", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16539 + }, + { + "path": "15317.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "dog", + "choice_c": "cat", + "choice_d": "fox", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16540 + }, + { + "path": "20015.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "Remote control car", + "choice_c": "remote controlled aircraft", + "choice_d": "The sound of electric saw", + "answer_gt": "remote controlled aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16541 + }, + { + "path": "58358.flac", + "question": "What is shown in the video?", + "choice_a": "Skateboarding and table tennis", + "choice_b": "play", + "choice_c": "Play in circles", + "choice_d": "Pick up the ball", + "answer_gt": "Skateboarding and table tennis", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16542 + }, + { + "path": "39434.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "bus", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16543 + }, + { + "path": "26217.flac", + "question": "Where did the video take place?", + "choice_a": "In the forest", + "choice_b": "playground", + "choice_c": "at sea", + "choice_d": "ski field", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16544 + }, + { + "path": "51817.flac", + "question": "Why does the chicken in the video cry?", + "choice_a": "It's daybreak", + "choice_b": "I want to find a company", + "choice_c": "Think of a cage", + "choice_d": "No worms", + "answer_gt": "Think of a cage", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16545 + }, + { + "path": "17669.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "wolf", + "choice_c": "rabbit", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16546 + }, + { + "path": "43738.flac", + "question": "What animal appears in the video?", + "choice_a": "wolf", + "choice_b": "Gibbon", + "choice_c": "black bear", + "choice_d": "macaque", + "answer_gt": "Gibbon", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16547 + }, + { + "path": "33669.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "sound of rain", + "choice_c": "thunder", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16548 + }, + { + "path": "45426.flac", + "question": "Where did the video take place?", + "choice_a": "toilet", + "choice_b": "forest", + "choice_c": "of one's own unit", + "choice_d": "courtyard", + "answer_gt": "of one's own unit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16549 + }, + { + "path": "57349.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Goat calls", + "choice_b": "The wolf barked", + "choice_c": "A pig crows", + "choice_d": "A donkey barks", + "answer_gt": "Goat calls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16550 + }, + { + "path": "4079.flac", + "question": "What happened in the video?", + "choice_a": "snow", + "choice_b": "It's Hailing", + "choice_c": "rain", + "choice_d": "Debris flow", + "answer_gt": "It's Hailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16551 + }, + { + "path": "53437.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Car horn", + "choice_b": "Engine acceleration", + "choice_c": "Ma Dasheng", + "choice_d": "Generator sound", + "answer_gt": "Engine acceleration", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16552 + }, + { + "path": "24404.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "cattle", + "choice_d": "donkey", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16553 + }, + { + "path": "51371.flac", + "question": "Why did the crow in the video fly?", + "choice_a": "I'm hungry", + "choice_b": "I was scared", + "choice_c": "it's raining", + "choice_d": "I hit it", + "answer_gt": "I was scared", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16554 + }, + { + "path": "59578.flac", + "question": "What is shown in the video?", + "choice_a": "pool", + "choice_b": "Bathtub", + "choice_c": "Shower", + "choice_d": "closestool", + "answer_gt": "closestool", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16555 + }, + { + "path": "17724.flac", + "question": "What animal appears in the video?", + "choice_a": "bird", + "choice_b": "kitty", + "choice_c": "chick", + "choice_d": "chipmunks", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16556 + }, + { + "path": "59729.flac", + "question": "What are the people doing in the video?", + "choice_a": "skating", + "choice_b": "Play balance car", + "choice_c": "Playing on the seesaw", + "choice_d": "Skateboarding", + "answer_gt": "Skateboarding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16557 + }, + { + "path": "44567.flac", + "question": "What animal appears in the video?", + "choice_a": "rowboat", + "choice_b": "elephant", + "choice_c": "Sea lions", + "choice_d": "Otter", + "answer_gt": "Sea lions", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16558 + }, + { + "path": "61096.flac", + "question": "What are the people doing in the video?", + "choice_a": "play table tennis", + "choice_b": "play volleyball", + "choice_c": "Play baseball", + "choice_d": "Play Cuju", + "answer_gt": "play volleyball", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16559 + }, + { + "path": "35954.flac", + "question": "What animal appears in the video?", + "choice_a": "duck", + "choice_b": "peacock", + "choice_c": "pheasant", + "choice_d": "tiger", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16560 + }, + { + "path": "37687.flac", + "question": "Where does the video take place?", + "choice_a": "water 's edge", + "choice_b": "playground", + "choice_c": "zoo", + "choice_d": "grassland", + "answer_gt": "water 's edge", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16561 + }, + { + "path": "25473.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Fire engine siren", + "choice_b": "police car siren", + "choice_c": "Ambulance siren", + "choice_d": "Civil defense alarm", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16562 + }, + { + "path": "29202.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "car engine noise", + "choice_c": "thunder", + "choice_d": "aircraft", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16563 + }, + { + "path": "13724.flac", + "question": "Where does the video take place?", + "choice_a": "aquatic", + "choice_b": "on the road", + "choice_c": "Grand Canyon", + "choice_d": "in the mountains", + "answer_gt": "in the mountains", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16564 + }, + { + "path": "5481.flac", + "question": "What are the people doing in the video?", + "choice_a": "Set off firecrackers", + "choice_b": "performing oral skills", + "choice_c": "Listen to the music", + "choice_d": "Shooting", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16565 + }, + { + "path": "36212.flac", + "question": "Where did the video take place?", + "choice_a": "highway", + "choice_b": "surface of the water", + "choice_c": "playground", + "choice_d": "at sea", + "answer_gt": "surface of the water", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16566 + }, + { + "path": "42584.flac", + "question": "Where is the bird in the video?", + "choice_a": "On the grass", + "choice_b": "cage", + "choice_c": "On the wire", + "choice_d": "On the branch", + "answer_gt": "On the branch", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16567 + }, + { + "path": "39371.flac", + "question": "What are the people in the video doing?", + "choice_a": "to shave", + "choice_b": "Machine gun fire", + "choice_c": "aim", + "choice_d": "go hunting", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16568 + }, + { + "path": "35552.flac", + "question": "What happened in the video?", + "choice_a": "The car is on fire", + "choice_b": "Set off firecrackers", + "choice_c": "Volcano eruption", + "choice_d": "blast", + "answer_gt": "The car is on fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16569 + }, + { + "path": "14536.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "automobile", + "choice_b": "tram", + "choice_c": "motorcycle", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16570 + }, + { + "path": "50239.flac", + "question": "What animal appears in the video?", + "choice_a": "snake", + "choice_b": "wolf", + "choice_c": "Loach", + "choice_d": "zebra", + "answer_gt": "snake", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16571 + }, + { + "path": "60867.flac", + "question": "Where does the video take place?", + "choice_a": "street", + "choice_b": "In the grassland", + "choice_c": "The sea", + "choice_d": "On the rocks", + "answer_gt": "The sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16572 + }, + { + "path": "16170.flac", + "question": "What are the people in the video doing?", + "choice_a": "Looking for something", + "choice_b": "Catching fish", + "choice_c": "diving", + "choice_d": "Breath holding exercises", + "answer_gt": "diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16573 + }, + { + "path": "37562.flac", + "question": "What happened in the video?", + "choice_a": "Volcanic explosion", + "choice_b": "tornado", + "choice_c": "firecrackers", + "choice_d": "volcanic explosion", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16574 + }, + { + "path": "30836.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "park", + "choice_c": "farm", + "choice_d": "Chicken coop", + "answer_gt": "Chicken coop", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16575 + }, + { + "path": "31894.flac", + "question": "What are the people doing in the video?", + "choice_a": "rubber band skipping", + "choice_b": "Remote control helicopter", + "choice_c": "fly a kite", + "choice_d": "Tug of war", + "answer_gt": "fly a kite", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16576 + }, + { + "path": "37583.flac", + "question": "Where does the video take place?", + "choice_a": "seaside", + "choice_b": "zoo", + "choice_c": "field", + "choice_d": "woods", + "answer_gt": "woods", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16577 + }, + { + "path": "37556.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "leopard", + "choice_c": "The bear", + "choice_d": "tiger", + "answer_gt": "tiger", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16578 + }, + { + "path": "46073.flac", + "question": "What animal appears in the video?", + "choice_a": "honeybee", + "choice_b": "crocodile", + "choice_c": "Magpie", + "choice_d": "wild goose", + "answer_gt": "Magpie", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16579 + }, + { + "path": "28539.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "horse", + "choice_c": "camel", + "choice_d": "rabbit", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16580 + }, + { + "path": "17811.flac", + "question": "Where did the video take place?", + "choice_a": "street", + "choice_b": "ski field", + "choice_c": "forest", + "choice_d": "at sea", + "answer_gt": "forest", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16581 + }, + { + "path": "14661.flac", + "question": "What are the people doing in the video?", + "choice_a": "racing", + "choice_b": "riding", + "choice_c": "Mountaineering", + "choice_d": "skiing", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16582 + }, + { + "path": "21843.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16583 + }, + { + "path": "53675.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Dog barking", + "choice_b": "The pigeon flapped its wings", + "choice_c": "The wind is blowing", + "choice_d": "cluck", + "answer_gt": "The pigeon flapped its wings", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16584 + }, + { + "path": "28430.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "motorcycle", + "choice_c": "motorboat", + "choice_d": "racing", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16585 + }, + { + "path": "39023.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "desert", + "choice_c": "field", + "choice_d": "zoo", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16586 + }, + { + "path": "27326.flac", + "question": "Where does the video take place?", + "choice_a": "playground", + "choice_b": "desert", + "choice_c": "Grassland", + "choice_d": "forest", + "answer_gt": "forest", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16587 + }, + { + "path": "48700.flac", + "question": "What are the people doing in the video?", + "choice_a": "Feed the chickens", + "choice_b": "Feed the cat", + "choice_c": "feed a dog", + "choice_d": "Feeding cattle", + "answer_gt": "feed a dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16588 + }, + { + "path": "7645.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "turkey", + "choice_c": "bird", + "choice_d": "cat", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16589 + }, + { + "path": "14792.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Fire engine siren", + "choice_b": "Civil defense alarm", + "choice_c": "smoke detector", + "choice_d": "Ambulance alert", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16590 + }, + { + "path": "43524.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "sliding door", + "choice_b": "gun", + "choice_c": "aircraft", + "choice_d": "truck", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16591 + }, + { + "path": "49800.flac", + "question": "Where did the video take place?", + "choice_a": "Lakeside", + "choice_b": "field", + "choice_c": "park", + "choice_d": "forest", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16592 + }, + { + "path": "57891.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Electric hairdresser", + "choice_b": "hair drier", + "choice_c": "Electric shaver", + "choice_d": "Massager", + "answer_gt": "Electric shaver", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16593 + }, + { + "path": "58625.flac", + "question": "What are the people doing in the video?", + "choice_a": "play golf", + "choice_b": "play tennis", + "choice_c": "shuttlecock", + "choice_d": "Play baseball", + "answer_gt": "play golf", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16594 + }, + { + "path": "57262.flac", + "question": "What happened in the video?", + "choice_a": "Sailing", + "choice_b": "Discovery of dolphins", + "choice_c": "Cargo ship running", + "choice_d": "Motorboat driving", + "answer_gt": "Sailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16595 + }, + { + "path": "28711.flac", + "question": "What is the main sound source of the video?", + "choice_a": "cat", + "choice_b": "dove", + "choice_c": "Toys", + "choice_d": "bird", + "answer_gt": "Toys", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16596 + }, + { + "path": "9987.flac", + "question": "What is the main sound source of the video?", + "choice_a": "hail", + "choice_b": "car engine noise", + "choice_c": "vacuum cleaner to clean the floor", + "choice_d": "Knock", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16597 + }, + { + "path": "21935.flac", + "question": "Where did the video take place?", + "choice_a": "On the subway", + "choice_b": "on board", + "choice_c": "On the plane", + "choice_d": "On the bus", + "answer_gt": "On the bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16598 + }, + { + "path": "39996.flac", + "question": "What animal is in the video?", + "choice_a": "cattle", + "choice_b": "duck", + "choice_c": "chicken", + "choice_d": "goose", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16599 + }, + { + "path": "35571.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "thunder", + "choice_c": "hail", + "choice_d": "The sound of the trumpet", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16600 + }, + { + "path": "46162.flac", + "question": "What sports are the people doing in the video?", + "choice_a": "high jump", + "choice_b": "Rope skipping", + "choice_c": "dance", + "choice_d": "long jump", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16601 + }, + { + "path": "23988.flac", + "question": "What animal is in the video?", + "choice_a": "bird", + "choice_b": "goose", + "choice_c": "chicken", + "choice_d": "duck", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16602 + }, + { + "path": "12369.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "cat", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16603 + }, + { + "path": "24423.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16604 + }, + { + "path": "44457.flac", + "question": "Where are the animals in the video?", + "choice_a": "cage", + "choice_b": "grassland", + "choice_c": "zoo", + "choice_d": "farm", + "answer_gt": "cage", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16605 + }, + { + "path": "32325.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "crow", + "choice_b": "dog's bark", + "choice_c": "Thrush calls", + "choice_d": "The Oriole calls", + "answer_gt": "Thrush calls", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16606 + }, + { + "path": "20798.flac", + "question": "What are the people doing in the video?", + "choice_a": "racing", + "choice_b": "riding", + "choice_c": "skiing", + "choice_d": "Riding a donkey", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16607 + }, + { + "path": "14758.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "wind chime", + "choice_b": "harmonica", + "choice_c": "Piano sound", + "choice_d": "Wind chime", + "answer_gt": "Wind chime", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16608 + }, + { + "path": "18171.flac", + "question": "What happened in the video?", + "choice_a": "pump", + "choice_b": "firecrackers", + "choice_c": "Volcanic explosion", + "choice_d": "volcanic explosion", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16609 + }, + { + "path": "54584.flac", + "question": "Where did the video take place?", + "choice_a": "stadium", + "choice_b": "Swimming Pool", + "choice_c": "theater", + "choice_d": "Dance room", + "answer_gt": "stadium", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16610 + }, + { + "path": "19036.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "car engine noise", + "choice_b": "hail", + "choice_c": "vacuum cleaner to clean the floor", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16611 + }, + { + "path": "53746.flac", + "question": "What are the people in the video doing?", + "choice_a": "Cutting cloth", + "choice_b": "Tearing paper", + "choice_c": "Folding paper crane", + "choice_d": "Shredding documents", + "answer_gt": "Tearing paper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16612 + }, + { + "path": "39389.flac", + "question": "What is the main sound source of the video?", + "choice_a": "sound of wind", + "choice_b": "sound of rain", + "choice_c": "thunder", + "choice_d": "Truck whistle", + "answer_gt": "thunder", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16613 + }, + { + "path": "42125.flac", + "question": "What are the people in the video doing?", + "choice_a": "drive a car", + "choice_b": "plough fields", + "choice_c": "Sweep the floor", + "choice_d": "cut grass", + "answer_gt": "cut grass", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16614 + }, + { + "path": "21116.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "dog", + "choice_d": "sheep", + "answer_gt": "sheep", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16615 + }, + { + "path": "3231.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "automobile", + "choice_b": "train", + "choice_c": "motorcycle", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16616 + }, + { + "path": "55808.flac", + "question": "What happened in the video?", + "choice_a": "Woodpeckers peck at trees", + "choice_b": "Parrots peck trees", + "choice_c": "Birds build nests", + "choice_d": "The parrot called", + "answer_gt": "Woodpeckers peck at trees", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16617 + }, + { + "path": "27027.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Smoke Alarm", + "choice_b": "Fire engine siren", + "choice_c": "honk", + "choice_d": "timer", + "answer_gt": "Smoke Alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16618 + }, + { + "path": "43620.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "train", + "choice_b": "bus", + "choice_c": "bus", + "choice_d": "Tractor", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16619 + }, + { + "path": "3371.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "frog", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "bird", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16620 + }, + { + "path": "55126.flac", + "question": "What are the people doing in the video?", + "choice_a": "Install alarm clock battery", + "choice_b": "Load a bullet", + "choice_c": "Installing mobile phone battery", + "choice_d": "Repair the remote control", + "answer_gt": "Install alarm clock battery", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16621 + }, + { + "path": "41099.flac", + "question": "What animal is in the video?", + "choice_a": "cattle", + "choice_b": "tiger", + "choice_c": "lion", + "choice_d": "leopard", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16622 + }, + { + "path": "4726.flac", + "question": "What are the people doing in the video?", + "choice_a": "performing oral skills", + "choice_b": "shooting practice", + "choice_c": "Shooting", + "choice_d": "Machine gun fire", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16623 + }, + { + "path": "22958.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16624 + }, + { + "path": "14040.flac", + "question": "What animal appears in the video?", + "choice_a": "frog", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "peacock", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16625 + }, + { + "path": "47507.flac", + "question": "Where is the bird in the video?", + "choice_a": "highway", + "choice_b": "farm", + "choice_c": "By the sea", + "choice_d": "grassland", + "answer_gt": "grassland", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16626 + }, + { + "path": "11093.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "metro", + "choice_c": "highway", + "choice_d": "aircraft", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16627 + }, + { + "path": "31228.flac", + "question": "What is the main sound source of the video?", + "choice_a": "thunder", + "choice_b": "Smoke Alarm", + "choice_c": "Civil defense alarm", + "choice_d": "Fire engine whistle", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16628 + }, + { + "path": "58768.flac", + "question": "What is shown in the video?", + "choice_a": "UAV", + "choice_b": "meteor shower", + "choice_c": "Fireworks and firecrackers", + "choice_d": "glowworm", + "answer_gt": "Fireworks and firecrackers", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16629 + }, + { + "path": "1308.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "chicken", + "choice_b": "duck", + "choice_c": "sparrow", + "choice_d": "dove", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16630 + }, + { + "path": "23456.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "sound of wind", + "choice_c": "aircraft", + "choice_d": "Machine sound", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16631 + }, + { + "path": "23577.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Drive a nail", + "choice_b": "screw", + "choice_c": "knock at the door", + "choice_d": "Knock on wood", + "answer_gt": "Drive a nail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16632 + }, + { + "path": "20573.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "elephant", + "choice_c": "tiger", + "choice_d": "lion", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16633 + }, + { + "path": "42701.flac", + "question": "What animal appears in the video?", + "choice_a": "chipmunks", + "choice_b": "parrot", + "choice_c": "dove", + "choice_d": "chicken", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16634 + }, + { + "path": "53133.flac", + "question": "What is shown in the video?", + "choice_a": "crane", + "choice_b": "Sliding door", + "choice_c": "Vertical ladder", + "choice_d": "Rolling door", + "answer_gt": "Sliding door", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16635 + }, + { + "path": "33112.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "firecrackers", + "choice_b": "Gun fire", + "choice_c": "blast", + "choice_d": "rope skipping", + "answer_gt": "Gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16636 + }, + { + "path": "10548.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16637 + }, + { + "path": "53842.flac", + "question": "What is the person in the video holding?", + "choice_a": "hair drier", + "choice_b": "Barber", + "choice_c": "razor", + "choice_d": "Scraping eyebrow knife", + "answer_gt": "razor", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16638 + }, + { + "path": "48916.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "sound of rain", + "choice_b": "thunder", + "choice_c": "voice", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16639 + }, + { + "path": "58532.flac", + "question": "What are the people doing in the video?", + "choice_a": "Feed the chickens", + "choice_b": "Catching birds", + "choice_c": "Catching chickens", + "choice_d": "Duck catching", + "answer_gt": "Catching chickens", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16640 + }, + { + "path": "47554.flac", + "question": "What's in the person's hand in the video?", + "choice_a": "to shave", + "choice_b": "wood stick", + "choice_c": "bowling", + "choice_d": "Pistol", + "answer_gt": "Pistol", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16641 + }, + { + "path": "59733.flac", + "question": "What is the main sound source in this video?", + "choice_a": "motor", + "choice_b": "Motorcycle engine noise", + "choice_c": "Car engine noise", + "choice_d": "Go kart engine", + "answer_gt": "Car engine noise", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16642 + }, + { + "path": "31878.flac", + "question": "What are the people doing in the video?", + "choice_a": "Swimming", + "choice_b": "rowboat", + "choice_c": "diving", + "choice_d": "go fishing", + "answer_gt": "rowboat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16643 + }, + { + "path": "28689.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorcycle", + "choice_b": "racing", + "choice_c": "motorboat", + "choice_d": "lawn mower", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16644 + }, + { + "path": "59584.flac", + "question": "What is shown in the video?", + "choice_a": "Electric welding", + "choice_b": "carving", + "choice_c": "printing", + "choice_d": "cutting", + "answer_gt": "carving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16645 + }, + { + "path": "2032.flac", + "question": "What animal is in the video?", + "choice_a": "crow", + "choice_b": "Magpie", + "choice_c": "parrot", + "choice_d": "Thrush", + "answer_gt": "crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16646 + }, + { + "path": "24998.flac", + "question": "What's in the person's hand in the video?", + "choice_a": "mobile phone", + "choice_b": "gun", + "choice_c": "lighter", + "choice_d": "bowling", + "answer_gt": "gun", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16647 + }, + { + "path": "25943.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "sheep", + "choice_d": "donkey", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16648 + }, + { + "path": "42180.flac", + "question": "What animal is in the video?", + "choice_a": "duck", + "choice_b": "rabbit", + "choice_c": "turkey", + "choice_d": "peacock", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16649 + }, + { + "path": "46789.flac", + "question": "What are the people doing in the video?", + "choice_a": "Wear a diving suit", + "choice_b": "wear glasses", + "choice_c": "put on shoes", + "choice_d": "disrobe", + "answer_gt": "Wear a diving suit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16650 + }, + { + "path": "27723.flac", + "question": "What happened in the video?", + "choice_a": "Reversing", + "choice_b": "light firecrackers", + "choice_c": "Repair the tire", + "choice_d": "Cars crush things", + "answer_gt": "Cars crush things", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16651 + }, + { + "path": "32619.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "automobile", + "choice_b": "Tractor", + "choice_c": "bus", + "choice_d": "motorcycle", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16652 + }, + { + "path": "1552.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "machine", + "choice_b": "remote controlled aircraft", + "choice_c": "motorboat", + "choice_d": "Tipper", + "answer_gt": "machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16653 + }, + { + "path": "23110.flac", + "question": "What happened in the video?", + "choice_a": "Water splashing", + "choice_b": "snow", + "choice_c": "take a shower", + "choice_d": "get wet in the rain", + "answer_gt": "Water splashing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16654 + }, + { + "path": "24023.flac", + "question": "What animal appears in the video?", + "choice_a": "duck", + "choice_b": "peacock", + "choice_c": "cock", + "choice_d": "elephant", + "answer_gt": "cock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16655 + }, + { + "path": "3026.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "waterfall", + "choice_b": "sea wave", + "choice_c": "Underwater sound", + "choice_d": "Streams", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16656 + }, + { + "path": "25686.flac", + "question": "Where does the video take place?", + "choice_a": "On the sea", + "choice_b": "racetrack", + "choice_c": "on the road", + "choice_d": "seabed", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16657 + }, + { + "path": "39438.flac", + "question": "What animal appears in the video?", + "choice_a": "leopard", + "choice_b": "lion", + "choice_c": "tiger", + "choice_d": "The bear", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16658 + }, + { + "path": "37423.flac", + "question": "What are the people in the video doing?", + "choice_a": "run", + "choice_b": "skating", + "choice_c": "skiing", + "choice_d": "dance", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16659 + }, + { + "path": "53898.flac", + "question": "What are the people doing in the video?", + "choice_a": "play tennis", + "choice_b": "play volleyball", + "choice_c": "shuttlecock", + "choice_d": "Play billiards", + "answer_gt": "play volleyball", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16660 + }, + { + "path": "38652.flac", + "question": "What is the main sound source of the video?", + "choice_a": "bus", + "choice_b": "bus", + "choice_c": "metro", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16661 + }, + { + "path": "56524.flac", + "question": "What happened in the video?", + "choice_a": "Car Wash", + "choice_b": "Oil change", + "choice_c": "Automobile engine knock", + "choice_d": "The car is leaking oil", + "answer_gt": "Automobile engine knock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16662 + }, + { + "path": "51934.flac", + "question": "Where is the cow in the video?", + "choice_a": "Grassland", + "choice_b": "zoo", + "choice_c": "Sheep farm", + "choice_d": "Cattle farm", + "answer_gt": "Cattle farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16663 + }, + { + "path": "25267.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "pig", + "choice_d": "cattle", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16664 + }, + { + "path": "53851.flac", + "question": "Where does the video take place?", + "choice_a": "Grassland", + "choice_b": "pigsty", + "choice_c": "Farmyard", + "choice_d": "farm", + "answer_gt": "farm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16665 + }, + { + "path": "60485.flac", + "question": "What are the people doing in the video?", + "choice_a": "Tearing paper", + "choice_b": "Newspaper gift", + "choice_c": "Reading newspapers", + "choice_d": "Folding paper", + "answer_gt": "Folding paper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16666 + }, + { + "path": "61209.flac", + "question": "What are the people doing in the video?", + "choice_a": "play table tennis", + "choice_b": "shuttlecock", + "choice_c": "Play baseball", + "choice_d": "Boxing", + "answer_gt": "play table tennis", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16667 + }, + { + "path": "25246.flac", + "question": "What are the people doing in the video?", + "choice_a": "run", + "choice_b": "Mountaineering", + "choice_c": "skiing", + "choice_d": "skating", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16668 + }, + { + "path": "54041.flac", + "question": "What are the people in the video doing?", + "choice_a": "go fishing", + "choice_b": "play tennis", + "choice_c": "play table tennis", + "choice_d": "play volleyball", + "answer_gt": "play table tennis", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16669 + }, + { + "path": "49012.flac", + "question": "What's the weather in the video?", + "choice_a": "overcast", + "choice_b": "a sunny day", + "choice_c": "Foggy days", + "choice_d": "rain", + "answer_gt": "rain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16670 + }, + { + "path": "56756.flac", + "question": "What are the people doing in the video?", + "choice_a": "Driving a car", + "choice_b": "Driving a tractor", + "choice_c": "Ride a motorcycle", + "choice_d": "Go kart", + "answer_gt": "Driving a car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16671 + }, + { + "path": "5066.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "motorcycle", + "choice_c": "metro", + "choice_d": "bus", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16672 + }, + { + "path": "25707.flac", + "question": "Where did the video take place?", + "choice_a": "ski field", + "choice_b": "of one's own unit", + "choice_c": "indoor", + "choice_d": "restroom", + "answer_gt": "of one's own unit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16673 + }, + { + "path": "19001.flac", + "question": "What is the main sound source of the video?", + "choice_a": "horn", + "choice_b": "mock bird", + "choice_c": "toy chicks", + "choice_d": "Playing flute", + "answer_gt": "toy chicks", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16674 + }, + { + "path": "34376.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Smoke Alarm", + "choice_b": "Fire engine siren", + "choice_c": "Train horn", + "choice_d": "The sound of playing the piano", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16675 + }, + { + "path": "35577.flac", + "question": "What happened in the video?", + "choice_a": "Motorcycle race", + "choice_b": "motorboat", + "choice_c": "Car race", + "choice_d": "Car race", + "answer_gt": "Motorcycle race", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16676 + }, + { + "path": "54255.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "The sound of the trumpet", + "choice_b": "The sound of motorcycles", + "choice_c": "engine", + "choice_d": "Water pump", + "answer_gt": "engine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16677 + }, + { + "path": "13387.flac", + "question": "Where did the video take place?", + "choice_a": "Playground", + "choice_b": "On the snow mountain", + "choice_c": "natatorium", + "choice_d": "Snow", + "answer_gt": "Snow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16678 + }, + { + "path": "12158.flac", + "question": "Where did the video take place?", + "choice_a": "at sea", + "choice_b": "seabed", + "choice_c": "on the road", + "choice_d": "On the street", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16679 + }, + { + "path": "26140.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "aircraft", + "choice_b": "metro", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16680 + }, + { + "path": "21968.flac", + "question": "What is the main sound source of the video?", + "choice_a": "remote controlled aircraft", + "choice_b": "gun", + "choice_c": "aircraft", + "choice_d": "vocal imitation performance", + "answer_gt": "aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16681 + }, + { + "path": "11896.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "hail", + "choice_c": "vacuum cleaner to clean the floor", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16682 + }, + { + "path": "24694.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Machine sound", + "choice_b": "sound of wind", + "choice_c": "motorcycle", + "choice_d": "acceleration sound of automobile engine", + "answer_gt": "Machine sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16683 + }, + { + "path": "21478.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "donkey", + "choice_d": "dog", + "answer_gt": "horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16684 + }, + { + "path": "4218.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "helicopter", + "choice_c": "ambulance", + "choice_d": "highway", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16685 + }, + { + "path": "6210.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Shooting", + "choice_b": "Fireworks display", + "choice_c": "blast", + "choice_d": "bombardment", + "answer_gt": "Fireworks display", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16686 + }, + { + "path": "4581.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "sheep", + "choice_c": "dog", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16687 + }, + { + "path": "58635.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Washing machine", + "choice_b": "Toilet flush sound", + "choice_c": "The sound of raindrops", + "choice_d": "Bath sound", + "answer_gt": "Toilet flush sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16688 + }, + { + "path": "35233.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16689 + }, + { + "path": "36074.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of wind", + "choice_b": "Toy car", + "choice_c": "Toy airplane", + "choice_d": "Toy train", + "answer_gt": "Toy airplane", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16690 + }, + { + "path": "23271.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "bird", + "choice_c": "wolf", + "choice_d": "cat", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16691 + }, + { + "path": "20930.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Civil defense alarm", + "choice_b": "Fire engine siren", + "choice_c": "Smoke Alarm", + "choice_d": "music", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16692 + }, + { + "path": "6963.flac", + "question": "What animal is in the video?", + "choice_a": "dove", + "choice_b": "chicken", + "choice_c": "owl", + "choice_d": "myna", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16693 + }, + { + "path": "21766.flac", + "question": "What animal is in the video?", + "choice_a": "tiger", + "choice_b": "lion", + "choice_c": "monkey", + "choice_d": "The bear", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16694 + }, + { + "path": "60989.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Car whistle", + "choice_b": "The sound of car driving", + "choice_c": "Sound of motorcycle running", + "choice_d": "The sound of ship engine", + "answer_gt": "The sound of car driving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16695 + }, + { + "path": "18810.flac", + "question": "Where does the video take place?", + "choice_a": "aquatic", + "choice_b": "street", + "choice_c": "of one's own unit", + "choice_d": "Out of hospital", + "answer_gt": "of one's own unit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16696 + }, + { + "path": "3047.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "deer", + "choice_c": "dog", + "choice_d": "fox", + "answer_gt": "fox", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16697 + }, + { + "path": "1299.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "remote controlled aircraft", + "choice_c": "aircraft", + "choice_d": "highway", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16698 + }, + { + "path": "42768.flac", + "question": "What animal is in the video?", + "choice_a": "chipmunks", + "choice_b": "goat", + "choice_c": "pig", + "choice_d": "cat", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16699 + }, + { + "path": "2868.flac", + "question": "What animal appears in the video?", + "choice_a": "cowbell", + "choice_b": "lion", + "choice_c": "elephant", + "choice_d": "rhinoceros", + "answer_gt": "elephant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16700 + }, + { + "path": "39125.flac", + "question": "What's the weather in the video?", + "choice_a": "It's Hailing", + "choice_b": "rain", + "choice_c": "a sunny day", + "choice_d": "overcast", + "answer_gt": "a sunny day", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16701 + }, + { + "path": "47943.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "Alpaca", + "choice_d": "pig", + "answer_gt": "donkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16702 + }, + { + "path": "44826.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Sheep crow", + "choice_b": "The horse barked", + "choice_c": "The donkey barked", + "choice_d": "The cow barked", + "answer_gt": "The cow barked", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16703 + }, + { + "path": "31924.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "sound of rain", + "choice_b": "Underwater sound", + "choice_c": "The sound of the stream", + "choice_d": "thunder", + "answer_gt": "The sound of the stream", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16704 + }, + { + "path": "10582.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Toy car", + "choice_b": "Toy horse", + "choice_c": "The toy dog", + "choice_d": "People imitate horses", + "answer_gt": "Toy horse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16705 + }, + { + "path": "55333.flac", + "question": "What is shown in the video?", + "choice_a": "motorcycle engine", + "choice_b": "ATV engine", + "choice_c": "Truck engine", + "choice_d": "automobile engine", + "answer_gt": "automobile engine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16706 + }, + { + "path": "54207.flac", + "question": "What are the people doing in the video?", + "choice_a": "disco dancing", + "choice_b": "speech", + "choice_c": "Performing oral skills", + "choice_d": "Performing and singing", + "answer_gt": "Performing oral skills", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16707 + }, + { + "path": "11841.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "fire engine", + "choice_b": "smoke detector", + "choice_c": "alarm clock", + "choice_d": "BB machine", + "answer_gt": "smoke detector", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16708 + }, + { + "path": "60452.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Fire fighting", + "choice_b": "Roasted sweet potato", + "choice_c": "The oven exploded", + "choice_d": "The crackle of a fire", + "answer_gt": "The crackle of a fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16709 + }, + { + "path": "33837.flac", + "question": "Where does the video take place?", + "choice_a": "cage", + "choice_b": "Amusement Park", + "choice_c": "Botanical Garden", + "choice_d": "zoo", + "answer_gt": "zoo", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16710 + }, + { + "path": "1514.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "automobile", + "choice_b": "horn", + "choice_c": "high-speed rail", + "choice_d": "train", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16711 + }, + { + "path": "12742.flac", + "question": "What animal appears in the video?", + "choice_a": "elephant", + "choice_b": "Gibbon", + "choice_c": "chimpanzee", + "choice_d": "macaque", + "answer_gt": "Gibbon", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16712 + }, + { + "path": "6293.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "automobile", + "choice_b": "metro", + "choice_c": "train", + "choice_d": "Tractor", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16713 + }, + { + "path": "41223.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "mouse", + "choice_c": "dog", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16714 + }, + { + "path": "33082.flac", + "question": "What animal is in the video?", + "choice_a": "fox", + "choice_b": "wolf", + "choice_c": "cat", + "choice_d": "rabbit", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16715 + }, + { + "path": "40441.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "sheep", + "choice_c": "fox", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16716 + }, + { + "path": "2285.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "automobile", + "choice_b": "metro", + "choice_c": "train", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16717 + }, + { + "path": "49038.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "horn", + "choice_b": "flute", + "choice_c": "Airplane noise", + "choice_d": "music", + "answer_gt": "music", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16718 + }, + { + "path": "17188.flac", + "question": "What are the people in the video doing?", + "choice_a": "Eat something", + "choice_b": "eat something", + "choice_c": "eat fruit", + "choice_d": "brush one 's teeth", + "answer_gt": "Eat something", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16719 + }, + { + "path": "12969.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "motorcycle", + "choice_c": "bus", + "choice_d": "metro", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16720 + }, + { + "path": "27835.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Fire engine siren", + "choice_b": "Police car siren", + "choice_c": "Smoke Alarm", + "choice_d": "Ambulance siren", + "answer_gt": "Police car siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16721 + }, + { + "path": "59272.flac", + "question": "What is shown in the video?", + "choice_a": "firecrackers", + "choice_b": "Fire fighting", + "choice_c": "Burn a fire", + "choice_d": "Boil water", + "answer_gt": "Burn a fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16722 + }, + { + "path": "31939.flac", + "question": "What are the people doing in the video?", + "choice_a": "play with water", + "choice_b": "long jump", + "choice_c": "Rope skipping", + "choice_d": "High jump", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16723 + }, + { + "path": "20499.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Fire engine siren", + "choice_b": "motorboat", + "choice_c": "Civil defense alarm", + "choice_d": "Ambulance siren", + "answer_gt": "Ambulance siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16724 + }, + { + "path": "39078.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "donkey", + "choice_c": "sheep", + "choice_d": "cattle", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16725 + }, + { + "path": "60028.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Motorcycle engine noise", + "choice_b": "Car engine noise", + "choice_c": "Sound of excavator engine", + "choice_d": "Snowmobile engine", + "answer_gt": "Car engine noise", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16726 + }, + { + "path": "20768.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "peacock", + "choice_c": "turkey", + "choice_d": "vulture", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16727 + }, + { + "path": "8322.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "bird", + "choice_b": "owl", + "choice_c": "duck", + "choice_d": "A rooster crows", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16728 + }, + { + "path": "54106.flac", + "question": "What happened in the video?", + "choice_a": "sing", + "choice_b": "vocal imitation performance", + "choice_c": "dance", + "choice_d": "play the guitar", + "answer_gt": "vocal imitation performance", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16729 + }, + { + "path": "34347.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "fox", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16730 + }, + { + "path": "9077.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "dog", + "choice_c": "lion", + "choice_d": "tiger", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16731 + }, + { + "path": "8152.flac", + "question": "What sounds in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "aircraft", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16732 + }, + { + "path": "21688.flac", + "question": "What are the people in the video doing?", + "choice_a": "Using a sewing machine", + "choice_b": "Making clothes", + "choice_c": "woodpeckers peck at trees", + "choice_d": "Using a dust collector", + "answer_gt": "Using a sewing machine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16733 + }, + { + "path": "41554.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "chipmunks", + "choice_c": "lion", + "choice_d": "a hippocampus", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16734 + }, + { + "path": "48528.flac", + "question": "What are the people in the video doing?", + "choice_a": "Swimming", + "choice_b": "go fishing", + "choice_c": "drive a car", + "choice_d": "rowing", + "answer_gt": "rowing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16735 + }, + { + "path": "26731.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "dog", + "choice_b": "fireworks and firecrackers", + "choice_c": "cricket", + "choice_d": "The sound of frogs", + "answer_gt": "cricket", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16736 + }, + { + "path": "27661.flac", + "question": "Where did the video take place?", + "choice_a": "Water park", + "choice_b": "Playground", + "choice_c": "playground", + "choice_d": "zoo", + "answer_gt": "Playground", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16737 + }, + { + "path": "46931.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "bus", + "choice_b": "train", + "choice_c": "aircraft", + "choice_d": "Tractor", + "answer_gt": "bus", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16738 + }, + { + "path": "46631.flac", + "question": "What animal appears in the video?", + "choice_a": "leopard", + "choice_b": "black bear", + "choice_c": "tiger", + "choice_d": "lion", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16739 + }, + { + "path": "21756.flac", + "question": "What is the main sound source of the video?", + "choice_a": "performing oral skills", + "choice_b": "blast", + "choice_c": "Machine gun fire", + "choice_d": "Set off firecrackers", + "answer_gt": "Machine gun fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16740 + }, + { + "path": "26982.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Remote control car", + "choice_b": "Remote control train", + "choice_c": "Toy Tractor", + "choice_d": "remote controlled aircraft", + "answer_gt": "remote controlled aircraft", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16741 + }, + { + "path": "56274.flac", + "question": "Where does the video take place?", + "choice_a": "room", + "choice_b": "Car bottom", + "choice_c": "Car roof", + "choice_d": "In the car", + "answer_gt": "In the car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16742 + }, + { + "path": "33690.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "motorcycle", + "choice_b": "train", + "choice_c": "motorboat", + "choice_d": "bus", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16743 + }, + { + "path": "32595.flac", + "question": "Where did the video take place?", + "choice_a": "playground", + "choice_b": "aquatic", + "choice_c": "desert", + "choice_d": "in the air", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16744 + }, + { + "path": "22499.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "chicken", + "choice_c": "duck", + "choice_d": "bird", + "answer_gt": "duck", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16745 + }, + { + "path": "34933.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "dog", + "choice_c": "sheep", + "choice_d": "wolf", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16746 + }, + { + "path": "22938.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "bullfight", + "choice_c": "A moment of silence", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16747 + }, + { + "path": "24612.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "racing", + "choice_c": "aircraft", + "choice_d": "motorcycle", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16748 + }, + { + "path": "27963.flac", + "question": "Where did the video take place?", + "choice_a": "in the water", + "choice_b": "On the road", + "choice_c": "at sea", + "choice_d": "beach", + "answer_gt": "at sea", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16749 + }, + { + "path": "24473.flac", + "question": "What happened in the video?", + "choice_a": "Volcanic explosion", + "choice_b": "firecrackers", + "choice_c": "cars crush things", + "choice_d": "volcanic explosion", + "answer_gt": "Volcanic explosion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16750 + }, + { + "path": "40921.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "crow", + "choice_c": "duck", + "choice_d": "chicken", + "answer_gt": "duck", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16751 + }, + { + "path": "24088.flac", + "question": "What are the people in the video doing?", + "choice_a": "skiing", + "choice_b": "bullfight", + "choice_c": "Riding a donkey", + "choice_d": "riding", + "answer_gt": "riding", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16752 + }, + { + "path": "38513.flac", + "question": "What animal appears in the video?", + "choice_a": "cattle", + "choice_b": "duck", + "choice_c": "peacock", + "choice_d": "turkey", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16753 + }, + { + "path": "16671.flac", + "question": "What are the people doing in the video?", + "choice_a": "play the piano", + "choice_b": "Ring the bell", + "choice_c": "play the guitar", + "choice_d": "Playing Guqin", + "answer_gt": "play the piano", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16754 + }, + { + "path": "51807.flac", + "question": "What are the people doing in the video?", + "choice_a": "speech", + "choice_b": "Feed cattle grass", + "choice_c": "Pull the cattle away", + "choice_d": "Selling cattle", + "answer_gt": "Pull the cattle away", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16755 + }, + { + "path": "28140.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorcycle", + "choice_b": "metro", + "choice_c": "racing", + "choice_d": "a storage battery car", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16756 + }, + { + "path": "57454.flac", + "question": "Where does this video take place?", + "choice_a": "square", + "choice_b": "In the courtyard", + "choice_c": "In the classroom", + "choice_d": "theater", + "answer_gt": "In the classroom", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16757 + }, + { + "path": "22657.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "car engine noise", + "choice_b": "acceleration sound of automobile engine", + "choice_c": "horn", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16758 + }, + { + "path": "1615.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "remote controlled aircraft", + "choice_d": "motorcycle", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16759 + }, + { + "path": "32164.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "The sound of motorcycles", + "choice_b": "motorboat", + "choice_c": "the sound of driving a snowmobile", + "choice_d": "lawn mower", + "answer_gt": "The sound of motorcycles", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16760 + }, + { + "path": "9506.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "crow", + "choice_c": "swallow", + "choice_d": "canary", + "answer_gt": "swallow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16761 + }, + { + "path": "19492.flac", + "question": "What animal appears in the video?", + "choice_a": "cat", + "choice_b": "wolf", + "choice_c": "rabbit", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16762 + }, + { + "path": "52784.flac", + "question": "What happened in the video?", + "choice_a": "Swallows build nests", + "choice_b": "Swallows foraging", + "choice_c": "Swallow drink water", + "choice_d": "Woodpecker pecks wood", + "answer_gt": "Woodpecker pecks wood", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16763 + }, + { + "path": "37536.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "donkey", + "choice_d": "cattle", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16764 + }, + { + "path": "22377.flac", + "question": "What's driving in the video?", + "choice_a": "train", + "choice_b": "automobile", + "choice_c": "Electric vehicle", + "choice_d": "ship", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16765 + }, + { + "path": "1872.flac", + "question": "What is the main sound source of the video?", + "choice_a": "police car", + "choice_b": "train", + "choice_c": "automobile", + "choice_d": "aircraft", + "answer_gt": "police car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16766 + }, + { + "path": "28123.flac", + "question": "Where does the video take place?", + "choice_a": "street", + "choice_b": "laboratory", + "choice_c": "school", + "choice_d": "of one's own unit", + "answer_gt": "of one's own unit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16767 + }, + { + "path": "36321.flac", + "question": "What animal is in the video?", + "choice_a": "dog", + "choice_b": "chipmunks", + "choice_c": "cat", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16768 + }, + { + "path": "39740.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "screw", + "choice_b": "Nailing", + "choice_c": "Take the nail", + "choice_d": "Press", + "answer_gt": "Nailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16769 + }, + { + "path": "9125.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "cattle", + "choice_c": "donkey", + "choice_d": "dog", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16770 + }, + { + "path": "12988.flac", + "question": "What are the people in the video doing?", + "choice_a": "racing", + "choice_b": "play a game", + "choice_c": "Watch tv", + "choice_d": "Cycling", + "answer_gt": "play a game", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16771 + }, + { + "path": "33428.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "aircraft", + "choice_b": "train", + "choice_c": "bus", + "choice_d": "high-speed rail", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16772 + }, + { + "path": "12034.flac", + "question": "Where did the video take place?", + "choice_a": "rivers", + "choice_b": "Underwater", + "choice_c": "sea surface", + "choice_d": "bathhouse", + "answer_gt": "Underwater", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16773 + }, + { + "path": "18348.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "cat", + "choice_c": "wolf", + "choice_d": "elephant", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16774 + }, + { + "path": "41107.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "thunder", + "choice_b": "Drumming", + "choice_c": "hail", + "choice_d": "The sound of rain", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16775 + }, + { + "path": "27551.flac", + "question": "Where did the video take place?", + "choice_a": "seaside", + "choice_b": "Grassland", + "choice_c": "desert", + "choice_d": "forest", + "answer_gt": "Grassland", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16776 + }, + { + "path": "31274.flac", + "question": "What animal appears in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "fox", + "choice_d": "rabbit", + "answer_gt": "fox", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16777 + }, + { + "path": "55901.flac", + "question": "What animal is in the video?", + "choice_a": "zebra", + "choice_b": "Leopard", + "choice_c": "giraffe", + "choice_d": "jackal", + "answer_gt": "Leopard", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16778 + }, + { + "path": "40498.flac", + "question": "What animal is in the video?", + "choice_a": "planing", + "choice_b": "honeybee", + "choice_c": "mosquito", + "choice_d": "butterfly", + "answer_gt": "honeybee", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16779 + }, + { + "path": "15646.flac", + "question": "Where did the video take place?", + "choice_a": "in the water", + "choice_b": "of one's own unit", + "choice_c": "In the desert", + "choice_d": "aquatic", + "answer_gt": "aquatic", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16780 + }, + { + "path": "20489.flac", + "question": "What animal appears in the video?", + "choice_a": "pheasant", + "choice_b": "dog", + "choice_c": "turkey", + "choice_d": "peacock", + "answer_gt": "pheasant", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16781 + }, + { + "path": "58072.flac", + "question": "What is shown in the video?", + "choice_a": "Electric sliding door", + "choice_b": "Electric windows", + "choice_c": "Warehouse door", + "choice_d": "Closet door", + "answer_gt": "Electric sliding door", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16782 + }, + { + "path": "13710.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "Ambulance siren", + "choice_b": "Fire engine siren", + "choice_c": "The train whistled", + "choice_d": "The police car whistled", + "answer_gt": "Fire engine siren", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16783 + }, + { + "path": "46653.flac", + "question": "What animal appears in the video?", + "choice_a": "honeybee", + "choice_b": "duck", + "choice_c": "earthworm", + "choice_d": "dove", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16784 + }, + { + "path": "40540.flac", + "question": "What animal is in the video?", + "choice_a": "goose", + "choice_b": "swallow", + "choice_c": "sparrow", + "choice_d": "crow", + "answer_gt": "swallow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16785 + }, + { + "path": "1705.flac", + "question": "What animal is in the video?", + "choice_a": "chicken", + "choice_b": "honeybee", + "choice_c": "Petrel", + "choice_d": "dove", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16786 + }, + { + "path": "53217.flac", + "question": "What is the purpose of the video?", + "choice_a": "Clean the toilet", + "choice_b": "Wash basin", + "choice_c": "Mop pool flush", + "choice_d": "Dredge the pipeline", + "answer_gt": "Clean the toilet", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16787 + }, + { + "path": "7941.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "car engine noise", + "choice_b": "voice", + "choice_c": "The sound of the stream", + "choice_d": "Underwater sound", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16788 + }, + { + "path": "11913.flac", + "question": "What are the people doing in the video?", + "choice_a": "racing", + "choice_b": "rowing", + "choice_c": "Rope skipping", + "choice_d": "printer printing", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16789 + }, + { + "path": "21055.flac", + "question": "What is the main sound source of the video?", + "choice_a": "remote controlled aircraft", + "choice_b": "hail", + "choice_c": "helicopter", + "choice_d": "acceleration sound of automobile engine", + "answer_gt": "helicopter", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16790 + }, + { + "path": "44403.flac", + "question": "Where does the video take place?", + "choice_a": "toilet", + "choice_b": "courtyard", + "choice_c": "zoo", + "choice_d": "of one's own unit", + "answer_gt": "of one's own unit", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16791 + }, + { + "path": "40513.flac", + "question": "Where does the video take place?", + "choice_a": "highway", + "choice_b": "toilet", + "choice_c": "On the viaduct", + "choice_d": "Mountain path", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16792 + }, + { + "path": "2157.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "wind", + "choice_b": "hail", + "choice_c": "scream", + "choice_d": "leaf", + "answer_gt": "hail", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16793 + }, + { + "path": "20734.flac", + "question": "What is the main sound source of the video?", + "choice_a": "motorcycle", + "choice_b": "horse racing", + "choice_c": "Tractor sound", + "choice_d": "racing", + "answer_gt": "racing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16794 + }, + { + "path": "23414.flac", + "question": "What animal is in the video?", + "choice_a": "owl", + "choice_b": "eagle", + "choice_c": "wild goose", + "choice_d": "turtledove", + "answer_gt": "owl", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16795 + }, + { + "path": "49371.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "dog's bark", + "choice_b": "A rooster crows", + "choice_c": "The donkey barked", + "choice_d": "Sheep crow", + "answer_gt": "Sheep crow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16796 + }, + { + "path": "48103.flac", + "question": "What is the sound source of the video?", + "choice_a": "Loach", + "choice_b": "Eel", + "choice_c": "snake", + "choice_d": "electric eel", + "answer_gt": "snake", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16797 + }, + { + "path": "12424.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "aircraft", + "choice_b": "motorboat", + "choice_c": "motorcycle", + "choice_d": "highway", + "answer_gt": "motorboat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16798 + }, + { + "path": "42272.flac", + "question": "Where did the video take place?", + "choice_a": "Lakeside", + "choice_b": "seabed", + "choice_c": "desert", + "choice_d": "seaside", + "answer_gt": "seaside", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16799 + }, + { + "path": "58021.flac", + "question": "What is shown in the video?", + "choice_a": "Electric windows", + "choice_b": "Electric curtain", + "choice_c": "Electric rolling shutter door", + "choice_d": "Electric transmission door", + "answer_gt": "Electric rolling shutter door", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16800 + }, + { + "path": "44435.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "dove", + "choice_b": "The cow barked", + "choice_c": "dog's bark", + "choice_d": "Cockcrow", + "answer_gt": "Cockcrow", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16801 + }, + { + "path": "29800.flac", + "question": "What happened in the video?", + "choice_a": "Cat and dog fight", + "choice_b": "The cat catches the mouse", + "choice_c": "The dog catches the rabbit", + "choice_d": "The cat teased the mouse", + "answer_gt": "The cat catches the mouse", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16802 + }, + { + "path": "49391.flac", + "question": "What's driving in the video?", + "choice_a": "Tractor", + "choice_b": "metro", + "choice_c": "bus", + "choice_d": "truck", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16803 + }, + { + "path": "37017.flac", + "question": "What is the main sound source of the video?", + "choice_a": "shot", + "choice_b": "rope skipping", + "choice_c": "firecrackers", + "choice_d": "The sound of lighters", + "answer_gt": "shot", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16804 + }, + { + "path": "11521.flac", + "question": "What are the people doing in this video?", + "choice_a": "Place a mask on the face", + "choice_b": "Make faces", + "choice_c": "Muscle movement", + "choice_d": "cry", + "answer_gt": "Make faces", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16805 + }, + { + "path": "38673.flac", + "question": "Where did the video take place?", + "choice_a": "Swimming Pool", + "choice_b": "In the pond", + "choice_c": "On the grass", + "choice_d": "In the water", + "answer_gt": "In the water", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16806 + }, + { + "path": "17935.flac", + "question": "What animal appears in the video?", + "choice_a": "dog", + "choice_b": "lion", + "choice_c": "cheetah", + "choice_d": "tiger", + "answer_gt": "lion", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16807 + }, + { + "path": "23184.flac", + "question": "What animal is in the video?", + "choice_a": "sheep", + "choice_b": "cat", + "choice_c": "wolf", + "choice_d": "dog", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16808 + }, + { + "path": "1503.flac", + "question": "What are the people doing in the video?", + "choice_a": "Rope skipping", + "choice_b": "rowing", + "choice_c": "applause", + "choice_d": "laugh", + "answer_gt": "Rope skipping", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16809 + }, + { + "path": "16054.flac", + "question": "What's running in the video?", + "choice_a": "train", + "choice_b": "Elevator.", + "choice_c": "aircraft", + "choice_d": "metro", + "answer_gt": "Elevator.", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16810 + }, + { + "path": "25757.flac", + "question": "Where did the video take place?", + "choice_a": "ski field", + "choice_b": "on the road", + "choice_c": "beach", + "choice_d": "seabed", + "answer_gt": "seabed", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16811 + }, + { + "path": "21838.flac", + "question": "Where did the video take place?", + "choice_a": "Ski Field", + "choice_b": "ski field", + "choice_c": "Amusement Park", + "choice_d": "skating rink", + "answer_gt": "Ski Field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16812 + }, + { + "path": "39351.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Machine sound", + "choice_b": "Car horn", + "choice_c": "pig", + "choice_d": "Wind chime", + "answer_gt": "Car horn", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16813 + }, + { + "path": "6290.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "sheep", + "choice_c": "cattle", + "choice_d": "camel", + "answer_gt": "cattle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16814 + }, + { + "path": "33352.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "goose", + "choice_c": "Oriole", + "choice_d": "Canary", + "answer_gt": "Canary", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16815 + }, + { + "path": "31535.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "rope skipping", + "choice_b": "thunder", + "choice_c": "The sound of fireworks", + "choice_d": "shot", + "answer_gt": "The sound of fireworks", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16816 + }, + { + "path": "6185.flac", + "question": "What animal is in the video?", + "choice_a": "horse", + "choice_b": "people", + "choice_c": "otter", + "choice_d": "donkey", + "answer_gt": "donkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16817 + }, + { + "path": "45715.flac", + "question": "What are the people doing in the video?", + "choice_a": "Play with dough", + "choice_b": "Play with crystal clay", + "choice_c": "Play with mud", + "choice_d": "Playing with sand", + "answer_gt": "Play with crystal clay", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16818 + }, + { + "path": "26754.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "train", + "choice_b": "Aircraft horn", + "choice_c": "Car horn", + "choice_d": "honk", + "answer_gt": "Car horn", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16819 + }, + { + "path": "5002.flac", + "question": "What's driving in the video?", + "choice_a": "fire engine", + "choice_b": "tram", + "choice_c": "automobile", + "choice_d": "bus", + "answer_gt": "automobile", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16820 + }, + { + "path": "38736.flac", + "question": "What is the main sound source of the video?", + "choice_a": "the sound of driving a snowmobile", + "choice_b": "bus", + "choice_c": "metro", + "choice_d": "motorcycle", + "answer_gt": "metro", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16821 + }, + { + "path": "53063.flac", + "question": "What are the people in the video doing?", + "choice_a": "Play with mud", + "choice_b": "Knead the dough", + "choice_c": "Play with crystal clay", + "choice_d": "Pinch soap", + "answer_gt": "Knead the dough", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16822 + }, + { + "path": "24970.flac", + "question": "What is the main sound source of the video?", + "choice_a": "dog", + "choice_b": "snake", + "choice_c": "earthworm", + "choice_d": "Loach", + "answer_gt": "snake", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16823 + }, + { + "path": "11977.flac", + "question": "What animal is in the video?", + "choice_a": "crocodile", + "choice_b": "snake", + "choice_c": "frog", + "choice_d": "Toad", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16824 + }, + { + "path": "56052.flac", + "question": "What are the people doing in the video?", + "choice_a": "play table tennis", + "choice_b": "play tennis", + "choice_c": "shuttlecock", + "choice_d": "Play billiards", + "answer_gt": "play table tennis", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16825 + }, + { + "path": "60740.flac", + "question": "What happened in the video?", + "choice_a": "Ship engine knock", + "choice_b": "Car dashboard turning", + "choice_c": "The engine is buzzing", + "choice_d": "Automobile engine knock", + "answer_gt": "Automobile engine knock", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16826 + }, + { + "path": "48012.flac", + "question": "What is the sound source of the video?", + "choice_a": "Geese crow", + "choice_b": "Quack", + "choice_c": "quack of a duck", + "choice_d": "A rooster crows", + "answer_gt": "A rooster crows", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16827 + }, + { + "path": "20045.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "duck", + "choice_c": "dog", + "choice_d": "peacock", + "answer_gt": "chicken", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16828 + }, + { + "path": "10943.flac", + "question": "What animal is in the video?", + "choice_a": "cat", + "choice_b": "dog", + "choice_c": "bird", + "choice_d": "pig", + "answer_gt": "dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16829 + }, + { + "path": "53729.flac", + "question": "What are the people in the video doing?", + "choice_a": "Acceleration tractor", + "choice_b": "Clean up rubbish", + "choice_c": "Mower mowing", + "choice_d": "Seeder seeding", + "answer_gt": "Acceleration tractor", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16830 + }, + { + "path": "48462.flac", + "question": "What are the main sources of sound in this video?", + "choice_a": "Tractor", + "choice_b": "bus", + "choice_c": "train", + "choice_d": "truck", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16831 + }, + { + "path": "40435.flac", + "question": "Where does the video take place?", + "choice_a": "toilet", + "choice_b": "In front of the mall", + "choice_c": "highway", + "choice_d": "at sea", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16832 + }, + { + "path": "4991.flac", + "question": "What is the main sound source of the video?", + "choice_a": "water", + "choice_b": "wind", + "choice_c": "hail", + "choice_d": "waterfall", + "answer_gt": "water", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16833 + }, + { + "path": "12950.flac", + "question": "What animal appears in the video?", + "choice_a": "horse", + "choice_b": "dog", + "choice_c": "cat", + "choice_d": "fox", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16834 + }, + { + "path": "29765.flac", + "question": "What animal is in the video?", + "choice_a": "rowboat", + "choice_b": "penguin", + "choice_c": "cockcrow", + "choice_d": "artillery", + "answer_gt": "penguin", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16835 + }, + { + "path": "3714.flac", + "question": "What are the people doing in the video?", + "choice_a": "Eat something", + "choice_b": "fishing", + "choice_c": "diving", + "choice_d": "bubble", + "answer_gt": "diving", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16836 + }, + { + "path": "27860.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of rain", + "choice_b": "sound of wind", + "choice_c": "thunder", + "choice_d": "zoo", + "answer_gt": "sound of rain", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16837 + }, + { + "path": "17390.flac", + "question": "What animal appears in the video?", + "choice_a": "kitty", + "choice_b": "bird", + "choice_c": "duck", + "choice_d": "chick", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16838 + }, + { + "path": "21599.flac", + "question": "What animal appears in the video?", + "choice_a": "kitty", + "choice_b": "turkey", + "choice_c": "dog", + "choice_d": "peacock", + "answer_gt": "turkey", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16839 + }, + { + "path": "18988.flac", + "question": "What happened in the video?", + "choice_a": "pump", + "choice_b": "volcanic explosion", + "choice_c": "It's Hailing", + "choice_d": "automobile engine starting", + "answer_gt": "It's Hailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16840 + }, + { + "path": "6708.flac", + "question": "What do the people in the video do?", + "choice_a": "telescope", + "choice_b": "skiing", + "choice_c": "Headgear", + "choice_d": "Remove the goggles", + "answer_gt": "Remove the goggles", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16841 + }, + { + "path": "1558.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "wind", + "choice_b": "people", + "choice_c": "water", + "choice_d": "highway", + "answer_gt": "people", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16842 + }, + { + "path": "2472.flac", + "question": "What is the main sound source of the video?", + "choice_a": "aircraft", + "choice_b": "wind", + "choice_c": "rain", + "choice_d": "water", + "answer_gt": "wind", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16843 + }, + { + "path": "13472.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "aircraft", + "choice_c": "train", + "choice_d": "motorcycle", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16844 + }, + { + "path": "54355.flac", + "question": "Where did the video take place?", + "choice_a": "stage", + "choice_b": "Small square", + "choice_c": "Concert Hall", + "choice_d": "On the street", + "answer_gt": "On the street", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16845 + }, + { + "path": "22453.flac", + "question": "Where does the video take place?", + "choice_a": "sound of wind", + "choice_b": "seabed", + "choice_c": "highway", + "choice_d": "cage", + "answer_gt": "highway", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16846 + }, + { + "path": "39935.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Birds", + "choice_b": "frog", + "choice_c": "Giant salamander", + "choice_d": "dragonfly", + "answer_gt": "frog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16847 + }, + { + "path": "57064.flac", + "question": "What happened in the video?", + "choice_a": "barbecue", + "choice_b": "Burn a fire", + "choice_c": "Fireworks", + "choice_d": "Firing charcoal", + "answer_gt": "Burn a fire", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16848 + }, + { + "path": "39926.flac", + "question": "What are the people doing in the video?", + "choice_a": "Eat something", + "choice_b": "drink water", + "choice_c": "train", + "choice_d": "Vomit", + "answer_gt": "Eat something", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16849 + }, + { + "path": "44869.flac", + "question": "Where did the video take place?", + "choice_a": "On the road", + "choice_b": "square", + "choice_c": "toilet", + "choice_d": "Grassland", + "answer_gt": "On the road", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16850 + }, + { + "path": "60491.flac", + "question": "What happened in the video?", + "choice_a": "Cargo ship sailing", + "choice_b": "Sailing accident", + "choice_c": "Sailing", + "choice_d": "Sail down", + "answer_gt": "Sailing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16851 + }, + { + "path": "19141.flac", + "question": "What's driving in the video?", + "choice_a": "automobile", + "choice_b": "tram", + "choice_c": "train", + "choice_d": "ship", + "answer_gt": "train", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16852 + }, + { + "path": "38861.flac", + "question": "Where did the video take place?", + "choice_a": "the river front", + "choice_b": "at sea", + "choice_c": "field", + "choice_d": "In the forest", + "answer_gt": "field", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16853 + }, + { + "path": "58782.flac", + "question": "What is shown in the video?", + "choice_a": "motorcycle engine", + "choice_b": "Car battery", + "choice_c": "The trunk of the car", + "choice_d": "automobile engine", + "answer_gt": "automobile engine", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16854 + }, + { + "path": "19579.flac", + "question": "What's in the person's hand in the video?", + "choice_a": "small bell", + "choice_b": "The ball", + "choice_c": "watermelon", + "choice_d": "Toys", + "answer_gt": "Toys", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16855 + }, + { + "path": "38691.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Underwater sound", + "choice_b": "thunder", + "choice_c": "firecrackers", + "choice_d": "sound of wind", + "answer_gt": "Underwater sound", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16856 + }, + { + "path": "25658.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "goose", + "choice_b": "bird", + "choice_c": "chicken", + "choice_d": "duck", + "answer_gt": "bird", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16857 + }, + { + "path": "37291.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "train", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "automobile", + "choice_d": "motorcycle", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16858 + }, + { + "path": "22819.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "Set off firecrackers", + "choice_b": "acceleration sound of automobile engine", + "choice_c": "performing oral skills", + "choice_d": "Shooting", + "answer_gt": "Shooting", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16859 + }, + { + "path": "17917.flac", + "question": "What animal appears in the video?", + "choice_a": "chicken", + "choice_b": "dove", + "choice_c": "dog", + "choice_d": "cockcrow", + "answer_gt": "dove", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16860 + }, + { + "path": "45201.flac", + "question": "What animal appears in the video?", + "choice_a": "the purr of a cat", + "choice_b": "cat", + "choice_c": "fox", + "choice_d": "elephant", + "answer_gt": "cat", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16861 + }, + { + "path": "35371.flac", + "question": "What are the people doing in the video?", + "choice_a": "Touch the dog", + "choice_b": "Give the dog a bath", + "choice_c": "Comb dog hair", + "choice_d": "Feed the dog", + "answer_gt": "Touch the dog", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16862 + }, + { + "path": "22785.flac", + "question": "What is the main sound source of the video?", + "choice_a": "train", + "choice_b": "police car", + "choice_c": "fire engine", + "choice_d": "aircraft", + "answer_gt": "police car", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16863 + }, + { + "path": "22665.flac", + "question": "What happened in the video?", + "choice_a": "Reversing", + "choice_b": "Cars crush things", + "choice_c": "light firecrackers", + "choice_d": "Repair the tire", + "answer_gt": "Cars crush things", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16864 + }, + { + "path": "54176.flac", + "question": "What are the people doing in the video?", + "choice_a": "paper-cut", + "choice_b": "Folding paper", + "choice_c": "Sticky paper", + "choice_d": "Tearing paper", + "answer_gt": "Tearing paper", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16865 + }, + { + "path": "24147.flac", + "question": "Where did the video take place?", + "choice_a": "ski field", + "choice_b": "at sea", + "choice_c": "forest", + "choice_d": "metro station", + "answer_gt": "forest", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16866 + }, + { + "path": "9066.flac", + "question": "What is the source of the sound in the video?", + "choice_a": "sound of wind", + "choice_b": "it's hailing", + "choice_c": "Civil defense alarm", + "choice_d": "erhu fiddle", + "answer_gt": "Civil defense alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16867 + }, + { + "path": "27212.flac", + "question": "What is the main source of sound in the video?", + "choice_a": "motorboat", + "choice_b": "the sound of driving a snowmobile", + "choice_c": "motorcycle", + "choice_d": "bus", + "answer_gt": "motorcycle", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16868 + }, + { + "path": "17885.flac", + "question": "What is the main sound source of the video?", + "choice_a": "Civil defense alarm", + "choice_b": "horn", + "choice_c": "computer", + "choice_d": "Smoke Alarm", + "answer_gt": "Smoke Alarm", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16869 + }, + { + "path": "24699.flac", + "question": "What are the people doing in the video?", + "choice_a": "run", + "choice_b": "Mountaineering", + "choice_c": "skiing", + "choice_d": "skating", + "answer_gt": "skiing", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16870 + }, + { + "path": "57943.flac", + "question": "What is the main sound source in this video?", + "choice_a": "Motorcycle engine noise", + "choice_b": "Car engine noise", + "choice_c": "Start sound of mower", + "choice_d": "Cruise engine noise", + "answer_gt": "Car engine noise", + "task_name": "Sound_AQA", + "dataset_name": "avqa", + "uniq_id": 16871 + }, + { + "path": "bass_synthetic_135-038-050.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16872 + }, + { + "path": "keyboard_electronic_098-044-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16873 + }, + { + "path": "bass_synthetic_033-041-050.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16874 + }, + { + "path": "keyboard_electronic_069-060-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16875 + }, + { + "path": "bass_synthetic_009-033-075.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16876 + }, + { + "path": "bass_synthetic_033-055-075.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16877 + }, + { + "path": "string_acoustic_071-051-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16878 + }, + { + "path": "vocal_synthetic_003-054-075.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16879 + }, + { + "path": "keyboard_electronic_098-043-075.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16880 + }, + { + "path": "reed_acoustic_023-054-100.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16881 + }, + { + "path": "reed_acoustic_023-044-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16882 + }, + { + "path": "keyboard_electronic_001-060-050.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16883 + }, + { + "path": "keyboard_acoustic_004-034-075.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16884 + }, + { + "path": "keyboard_electronic_098-102-100.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16885 + }, + { + "path": "reed_acoustic_023-055-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16886 + }, + { + "path": "guitar_acoustic_015-058-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16887 + }, + { + "path": "guitar_acoustic_015-043-127.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16888 + }, + { + "path": "bass_synthetic_009-025-050.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16889 + }, + { + "path": "organ_electronic_007-080-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16890 + }, + { + "path": "organ_electronic_001-053-127.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16891 + }, + { + "path": "string_acoustic_057-076-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16892 + }, + { + "path": "keyboard_synthetic_000-079-050.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16893 + }, + { + "path": "keyboard_electronic_001-069-050.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16894 + }, + { + "path": "organ_electronic_007-076-127.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16895 + }, + { + "path": "keyboard_acoustic_004-091-127.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16896 + }, + { + "path": "organ_electronic_028-094-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16897 + }, + { + "path": "keyboard_electronic_001-103-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16898 + }, + { + "path": "mallet_acoustic_062-075-075.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16899 + }, + { + "path": "string_acoustic_014-045-050.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16900 + }, + { + "path": "vocal_synthetic_003-048-100.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16901 + }, + { + "path": "keyboard_electronic_098-081-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16902 + }, + { + "path": "mallet_acoustic_062-053-075.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16903 + }, + { + "path": "reed_acoustic_023-039-025.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16904 + }, + { + "path": "bass_synthetic_009-045-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16905 + }, + { + "path": "flute_synthetic_000-071-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16906 + }, + { + "path": "guitar_acoustic_015-102-075.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16907 + }, + { + "path": "keyboard_synthetic_000-074-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16908 + }, + { + "path": "organ_electronic_007-048-025.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16909 + }, + { + "path": "bass_synthetic_134-059-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16910 + }, + { + "path": "guitar_electronic_022-059-025.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16911 + }, + { + "path": "bass_synthetic_134-060-075.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16912 + }, + { + "path": "keyboard_synthetic_000-081-100.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16913 + }, + { + "path": "bass_synthetic_009-044-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16914 + }, + { + "path": "reed_acoustic_037-075-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16915 + }, + { + "path": "guitar_acoustic_030-064-025.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16916 + }, + { + "path": "mallet_acoustic_062-083-025.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16917 + }, + { + "path": "flute_synthetic_000-026-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16918 + }, + { + "path": "keyboard_electronic_002-084-127.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16919 + }, + { + "path": "keyboard_electronic_078-046-127.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16920 + }, + { + "path": "reed_acoustic_023-086-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16921 + }, + { + "path": "keyboard_electronic_098-086-050.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16922 + }, + { + "path": "brass_acoustic_006-070-100.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16923 + }, + { + "path": "keyboard_electronic_002-083-075.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16924 + }, + { + "path": "string_acoustic_014-035-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16925 + }, + { + "path": "keyboard_electronic_001-049-075.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16926 + }, + { + "path": "keyboard_synthetic_000-104-100.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16927 + }, + { + "path": "brass_acoustic_046-089-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16928 + }, + { + "path": "string_acoustic_014-041-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16929 + }, + { + "path": "brass_acoustic_016-058-050.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "flute", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16930 + }, + { + "path": "guitar_acoustic_015-067-050.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16931 + }, + { + "path": "vocal_synthetic_003-081-075.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16932 + }, + { + "path": "keyboard_acoustic_004-098-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16933 + }, + { + "path": "organ_electronic_057-034-100.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16934 + }, + { + "path": "mallet_acoustic_056-088-025.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16935 + }, + { + "path": "string_acoustic_014-036-100.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16936 + }, + { + "path": "brass_acoustic_059-035-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16937 + }, + { + "path": "keyboard_electronic_098-103-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16938 + }, + { + "path": "flute_acoustic_002-095-100.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16939 + }, + { + "path": "keyboard_electronic_003-037-050.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16940 + }, + { + "path": "keyboard_acoustic_004-058-075.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16941 + }, + { + "path": "guitar_acoustic_030-062-127.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16942 + }, + { + "path": "keyboard_acoustic_004-067-025.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16943 + }, + { + "path": "bass_synthetic_098-040-025.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16944 + }, + { + "path": "brass_acoustic_006-073-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16945 + }, + { + "path": "organ_electronic_007-082-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16946 + }, + { + "path": "bass_synthetic_098-061-127.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16947 + }, + { + "path": "string_acoustic_014-062-127.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16948 + }, + { + "path": "keyboard_synthetic_000-051-100.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16949 + }, + { + "path": "mallet_acoustic_047-065-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16950 + }, + { + "path": "bass_electronic_027-022-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16951 + }, + { + "path": "organ_electronic_007-015-075.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16952 + }, + { + "path": "flute_synthetic_000-104-075.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16953 + }, + { + "path": "organ_electronic_007-030-050.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16954 + }, + { + "path": "keyboard_electronic_001-031-025.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16955 + }, + { + "path": "bass_synthetic_135-046-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16956 + }, + { + "path": "reed_acoustic_037-066-100.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16957 + }, + { + "path": "brass_acoustic_006-077-100.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16958 + }, + { + "path": "keyboard_electronic_078-038-100.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16959 + }, + { + "path": "reed_acoustic_023-042-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "vocal", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16960 + }, + { + "path": "mallet_acoustic_062-072-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16961 + }, + { + "path": "keyboard_synthetic_000-033-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16962 + }, + { + "path": "bass_synthetic_009-047-100.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16963 + }, + { + "path": "bass_synthetic_135-039-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16964 + }, + { + "path": "keyboard_synthetic_000-093-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16965 + }, + { + "path": "organ_electronic_007-011-050.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16966 + }, + { + "path": "bass_synthetic_033-058-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16967 + }, + { + "path": "guitar_electronic_022-040-100.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16968 + }, + { + "path": "organ_electronic_001-050-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16969 + }, + { + "path": "keyboard_electronic_002-091-025.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16970 + }, + { + "path": "bass_synthetic_135-044-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16971 + }, + { + "path": "bass_synthetic_033-049-075.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16972 + }, + { + "path": "organ_electronic_007-055-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16973 + }, + { + "path": "keyboard_electronic_069-078-025.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16974 + }, + { + "path": "bass_synthetic_135-043-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16975 + }, + { + "path": "bass_synthetic_135-027-025.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16976 + }, + { + "path": "keyboard_acoustic_004-104-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16977 + }, + { + "path": "bass_synthetic_135-073-127.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16978 + }, + { + "path": "keyboard_synthetic_000-025-100.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16979 + }, + { + "path": "organ_electronic_113-061-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16980 + }, + { + "path": "keyboard_electronic_078-040-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16981 + }, + { + "path": "guitar_acoustic_021-082-050.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16982 + }, + { + "path": "mallet_acoustic_062-098-025.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16983 + }, + { + "path": "bass_synthetic_134-050-050.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16984 + }, + { + "path": "keyboard_acoustic_004-085-075.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16985 + }, + { + "path": "bass_synthetic_134-076-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16986 + }, + { + "path": "bass_synthetic_134-055-100.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16987 + }, + { + "path": "guitar_acoustic_015-027-050.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16988 + }, + { + "path": "mallet_acoustic_047-066-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16989 + }, + { + "path": "keyboard_electronic_078-044-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16990 + }, + { + "path": "flute_synthetic_000-074-127.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16991 + }, + { + "path": "bass_synthetic_135-050-025.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16992 + }, + { + "path": "mallet_acoustic_056-035-050.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16993 + }, + { + "path": "keyboard_electronic_069-059-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16994 + }, + { + "path": "bass_synthetic_034-071-075.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16995 + }, + { + "path": "keyboard_synthetic_000-063-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16996 + }, + { + "path": "guitar_electronic_028-056-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16997 + }, + { + "path": "guitar_acoustic_014-100-075.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16998 + }, + { + "path": "brass_acoustic_006-025-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 16999 + }, + { + "path": "guitar_acoustic_014-099-075.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17000 + }, + { + "path": "organ_electronic_007-037-100.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17001 + }, + { + "path": "flute_acoustic_002-074-127.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17002 + }, + { + "path": "bass_synthetic_134-096-100.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17003 + }, + { + "path": "bass_synthetic_134-095-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17004 + }, + { + "path": "organ_electronic_007-020-025.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17005 + }, + { + "path": "keyboard_acoustic_004-103-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17006 + }, + { + "path": "bass_synthetic_068-037-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17007 + }, + { + "path": "mallet_acoustic_047-066-100.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17008 + }, + { + "path": "organ_electronic_028-068-075.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17009 + }, + { + "path": "keyboard_electronic_001-065-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17010 + }, + { + "path": "string_acoustic_056-075-100.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17011 + }, + { + "path": "flute_acoustic_002-095-025.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17012 + }, + { + "path": "keyboard_acoustic_004-069-127.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17013 + }, + { + "path": "keyboard_electronic_001-107-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17014 + }, + { + "path": "organ_electronic_057-040-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17015 + }, + { + "path": "keyboard_electronic_002-046-127.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17016 + }, + { + "path": "flute_synthetic_000-032-075.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17017 + }, + { + "path": "bass_synthetic_134-059-050.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17018 + }, + { + "path": "bass_synthetic_134-045-100.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17019 + }, + { + "path": "mallet_acoustic_062-085-050.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17020 + }, + { + "path": "mallet_acoustic_047-103-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17021 + }, + { + "path": "keyboard_acoustic_004-065-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17022 + }, + { + "path": "bass_synthetic_034-034-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17023 + }, + { + "path": "keyboard_synthetic_000-026-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17024 + }, + { + "path": "bass_synthetic_134-026-127.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17025 + }, + { + "path": "brass_acoustic_015-050-050.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17026 + }, + { + "path": "string_acoustic_014-029-100.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17027 + }, + { + "path": "mallet_acoustic_062-061-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17028 + }, + { + "path": "bass_synthetic_134-029-075.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17029 + }, + { + "path": "guitar_acoustic_021-025-127.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17030 + }, + { + "path": "organ_electronic_057-027-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17031 + }, + { + "path": "keyboard_electronic_002-104-075.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17032 + }, + { + "path": "guitar_electronic_028-106-075.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17033 + }, + { + "path": "bass_electronic_018-022-050.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17034 + }, + { + "path": "keyboard_electronic_002-059-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17035 + }, + { + "path": "keyboard_acoustic_004-023-075.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17036 + }, + { + "path": "string_acoustic_071-039-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17037 + }, + { + "path": "guitar_acoustic_015-098-050.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17038 + }, + { + "path": "brass_acoustic_006-078-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17039 + }, + { + "path": "flute_synthetic_000-101-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17040 + }, + { + "path": "reed_acoustic_023-037-025.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17041 + }, + { + "path": "guitar_acoustic_014-107-025.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17042 + }, + { + "path": "vocal_synthetic_003-055-075.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17043 + }, + { + "path": "keyboard_electronic_078-060-050.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17044 + }, + { + "path": "guitar_electronic_022-047-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17045 + }, + { + "path": "organ_electronic_113-041-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17046 + }, + { + "path": "keyboard_synthetic_000-093-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17047 + }, + { + "path": "keyboard_electronic_098-089-025.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17048 + }, + { + "path": "organ_electronic_007-046-100.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17049 + }, + { + "path": "keyboard_electronic_069-078-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17050 + }, + { + "path": "keyboard_acoustic_004-084-025.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17051 + }, + { + "path": "guitar_acoustic_030-079-127.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17052 + }, + { + "path": "guitar_acoustic_015-101-050.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17053 + }, + { + "path": "string_acoustic_071-038-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17054 + }, + { + "path": "bass_synthetic_135-051-050.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17055 + }, + { + "path": "keyboard_electronic_003-023-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17056 + }, + { + "path": "bass_synthetic_009-087-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17057 + }, + { + "path": "keyboard_acoustic_004-031-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17058 + }, + { + "path": "keyboard_acoustic_004-050-025.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17059 + }, + { + "path": "string_acoustic_012-026-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17060 + }, + { + "path": "organ_electronic_001-040-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17061 + }, + { + "path": "mallet_acoustic_047-085-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "flute", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17062 + }, + { + "path": "reed_acoustic_037-057-100.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17063 + }, + { + "path": "keyboard_electronic_098-094-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17064 + }, + { + "path": "organ_electronic_007-073-025.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17065 + }, + { + "path": "bass_synthetic_068-034-050.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17066 + }, + { + "path": "keyboard_electronic_098-054-025.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17067 + }, + { + "path": "keyboard_synthetic_000-077-025.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17068 + }, + { + "path": "organ_electronic_028-031-075.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17069 + }, + { + "path": "organ_electronic_113-033-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17070 + }, + { + "path": "guitar_acoustic_015-027-025.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17071 + }, + { + "path": "string_acoustic_057-073-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17072 + }, + { + "path": "brass_acoustic_016-081-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17073 + }, + { + "path": "vocal_synthetic_003-065-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17074 + }, + { + "path": "mallet_acoustic_062-059-075.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17075 + }, + { + "path": "guitar_electronic_028-033-100.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17076 + }, + { + "path": "string_acoustic_012-042-100.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17077 + }, + { + "path": "flute_synthetic_000-037-127.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17078 + }, + { + "path": "bass_synthetic_009-068-025.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17079 + }, + { + "path": "bass_synthetic_068-079-075.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17080 + }, + { + "path": "vocal_synthetic_003-106-050.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17081 + }, + { + "path": "flute_acoustic_002-082-075.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17082 + }, + { + "path": "keyboard_acoustic_004-051-050.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17083 + }, + { + "path": "mallet_acoustic_056-043-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17084 + }, + { + "path": "brass_acoustic_016-080-100.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17085 + }, + { + "path": "brass_acoustic_015-076-075.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17086 + }, + { + "path": "organ_electronic_007-017-127.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17087 + }, + { + "path": "reed_acoustic_023-076-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17088 + }, + { + "path": "flute_acoustic_002-090-127.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17089 + }, + { + "path": "keyboard_electronic_069-049-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17090 + }, + { + "path": "flute_synthetic_000-057-100.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17091 + }, + { + "path": "guitar_acoustic_030-021-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17092 + }, + { + "path": "guitar_acoustic_015-041-025.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17093 + }, + { + "path": "keyboard_acoustic_004-041-050.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17094 + }, + { + "path": "keyboard_acoustic_004-048-025.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17095 + }, + { + "path": "guitar_acoustic_015-098-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17096 + }, + { + "path": "keyboard_synthetic_000-076-025.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17097 + }, + { + "path": "guitar_acoustic_014-092-127.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17098 + }, + { + "path": "mallet_acoustic_047-103-075.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17099 + }, + { + "path": "organ_electronic_001-062-100.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17100 + }, + { + "path": "bass_synthetic_068-025-075.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17101 + }, + { + "path": "keyboard_acoustic_004-024-100.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17102 + }, + { + "path": "mallet_acoustic_056-075-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17103 + }, + { + "path": "keyboard_synthetic_000-039-075.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17104 + }, + { + "path": "keyboard_synthetic_000-076-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17105 + }, + { + "path": "bass_synthetic_034-052-100.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17106 + }, + { + "path": "guitar_acoustic_015-106-127.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17107 + }, + { + "path": "bass_synthetic_135-025-075.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17108 + }, + { + "path": "bass_synthetic_034-021-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17109 + }, + { + "path": "guitar_acoustic_015-059-050.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17110 + }, + { + "path": "mallet_acoustic_062-051-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17111 + }, + { + "path": "mallet_acoustic_062-050-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17112 + }, + { + "path": "vocal_synthetic_003-075-025.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17113 + }, + { + "path": "bass_synthetic_009-021-050.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17114 + }, + { + "path": "reed_acoustic_023-035-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17115 + }, + { + "path": "guitar_electronic_022-057-025.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17116 + }, + { + "path": "bass_synthetic_034-022-025.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17117 + }, + { + "path": "organ_electronic_028-105-127.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17118 + }, + { + "path": "organ_electronic_007-080-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17119 + }, + { + "path": "organ_electronic_007-078-050.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17120 + }, + { + "path": "keyboard_electronic_069-045-100.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17121 + }, + { + "path": "vocal_synthetic_003-036-025.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17122 + }, + { + "path": "reed_acoustic_037-068-050.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17123 + }, + { + "path": "bass_synthetic_034-037-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17124 + }, + { + "path": "brass_acoustic_016-068-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17125 + }, + { + "path": "keyboard_acoustic_004-079-025.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17126 + }, + { + "path": "keyboard_synthetic_000-095-127.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17127 + }, + { + "path": "guitar_electronic_022-056-100.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17128 + }, + { + "path": "guitar_electronic_022-047-127.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17129 + }, + { + "path": "guitar_acoustic_015-044-100.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17130 + }, + { + "path": "brass_acoustic_016-061-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17131 + }, + { + "path": "mallet_acoustic_062-051-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17132 + }, + { + "path": "organ_electronic_028-065-127.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17133 + }, + { + "path": "brass_acoustic_015-077-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17134 + }, + { + "path": "bass_synthetic_135-049-050.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17135 + }, + { + "path": "brass_acoustic_046-095-025.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17136 + }, + { + "path": "string_acoustic_056-058-025.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17137 + }, + { + "path": "reed_acoustic_037-048-050.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17138 + }, + { + "path": "bass_synthetic_009-019-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17139 + }, + { + "path": "keyboard_acoustic_004-023-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17140 + }, + { + "path": "guitar_acoustic_015-104-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17141 + }, + { + "path": "brass_acoustic_015-054-127.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17142 + }, + { + "path": "bass_synthetic_134-054-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17143 + }, + { + "path": "keyboard_electronic_001-085-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17144 + }, + { + "path": "organ_electronic_028-025-050.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17145 + }, + { + "path": "flute_acoustic_002-103-127.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17146 + }, + { + "path": "bass_synthetic_068-049-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17147 + }, + { + "path": "brass_acoustic_006-068-127.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17148 + }, + { + "path": "guitar_acoustic_030-097-127.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17149 + }, + { + "path": "mallet_acoustic_062-075-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17150 + }, + { + "path": "guitar_acoustic_021-075-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17151 + }, + { + "path": "keyboard_acoustic_004-023-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17152 + }, + { + "path": "bass_synthetic_068-024-025.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17153 + }, + { + "path": "mallet_acoustic_062-046-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17154 + }, + { + "path": "string_acoustic_080-039-050.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17155 + }, + { + "path": "reed_acoustic_037-034-127.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17156 + }, + { + "path": "string_acoustic_014-065-050.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17157 + }, + { + "path": "guitar_acoustic_015-094-127.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17158 + }, + { + "path": "guitar_acoustic_021-076-127.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17159 + }, + { + "path": "organ_electronic_113-029-127.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17160 + }, + { + "path": "flute_synthetic_000-077-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "vocal", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17161 + }, + { + "path": "bass_synthetic_009-096-025.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17162 + }, + { + "path": "brass_acoustic_046-092-050.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17163 + }, + { + "path": "string_acoustic_056-066-050.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17164 + }, + { + "path": "string_acoustic_014-051-127.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17165 + }, + { + "path": "organ_electronic_007-041-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17166 + }, + { + "path": "brass_acoustic_016-066-075.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17167 + }, + { + "path": "organ_electronic_007-064-100.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17168 + }, + { + "path": "keyboard_electronic_001-085-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17169 + }, + { + "path": "bass_synthetic_068-105-075.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17170 + }, + { + "path": "bass_synthetic_034-046-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17171 + }, + { + "path": "flute_acoustic_002-083-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17172 + }, + { + "path": "guitar_acoustic_014-088-050.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17173 + }, + { + "path": "flute_synthetic_000-056-050.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17174 + }, + { + "path": "string_acoustic_014-045-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17175 + }, + { + "path": "brass_acoustic_006-034-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17176 + }, + { + "path": "mallet_acoustic_062-051-050.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17177 + }, + { + "path": "brass_acoustic_015-045-075.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17178 + }, + { + "path": "keyboard_electronic_078-049-075.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17179 + }, + { + "path": "brass_acoustic_016-084-075.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17180 + }, + { + "path": "vocal_synthetic_003-054-050.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17181 + }, + { + "path": "brass_acoustic_046-071-100.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17182 + }, + { + "path": "mallet_acoustic_047-105-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17183 + }, + { + "path": "mallet_acoustic_062-062-025.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17184 + }, + { + "path": "organ_electronic_007-051-127.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17185 + }, + { + "path": "reed_acoustic_037-069-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17186 + }, + { + "path": "vocal_synthetic_003-085-127.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17187 + }, + { + "path": "reed_acoustic_037-037-127.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17188 + }, + { + "path": "organ_electronic_028-063-100.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17189 + }, + { + "path": "bass_synthetic_068-027-127.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17190 + }, + { + "path": "string_acoustic_056-078-100.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17191 + }, + { + "path": "organ_electronic_028-038-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17192 + }, + { + "path": "organ_electronic_057-052-050.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17193 + }, + { + "path": "vocal_synthetic_003-069-025.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17194 + }, + { + "path": "keyboard_electronic_098-042-025.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17195 + }, + { + "path": "guitar_acoustic_014-096-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17196 + }, + { + "path": "organ_electronic_007-086-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17197 + }, + { + "path": "organ_electronic_007-034-075.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17198 + }, + { + "path": "string_acoustic_080-038-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17199 + }, + { + "path": "guitar_acoustic_030-094-025.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17200 + }, + { + "path": "vocal_synthetic_003-103-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17201 + }, + { + "path": "reed_acoustic_023-071-127.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17202 + }, + { + "path": "flute_synthetic_000-105-100.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "reed", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17203 + }, + { + "path": "bass_synthetic_098-028-050.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "vocal", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17204 + }, + { + "path": "organ_electronic_057-031-127.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17205 + }, + { + "path": "guitar_electronic_028-029-075.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17206 + }, + { + "path": "organ_electronic_057-025-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "reed", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17207 + }, + { + "path": "mallet_acoustic_047-078-127.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17208 + }, + { + "path": "organ_electronic_028-061-050.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17209 + }, + { + "path": "keyboard_synthetic_000-026-050.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17210 + }, + { + "path": "reed_acoustic_037-073-025.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17211 + }, + { + "path": "brass_acoustic_006-054-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17212 + }, + { + "path": "mallet_acoustic_062-028-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17213 + }, + { + "path": "keyboard_electronic_078-034-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17214 + }, + { + "path": "organ_electronic_057-036-100.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17215 + }, + { + "path": "string_acoustic_056-039-025.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17216 + }, + { + "path": "reed_acoustic_011-053-100.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17217 + }, + { + "path": "string_acoustic_012-051-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17218 + }, + { + "path": "guitar_electronic_028-045-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17219 + }, + { + "path": "mallet_acoustic_056-058-050.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17220 + }, + { + "path": "mallet_acoustic_062-070-025.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17221 + }, + { + "path": "reed_acoustic_023-033-050.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17222 + }, + { + "path": "bass_synthetic_135-047-127.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17223 + }, + { + "path": "bass_electronic_018-062-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17224 + }, + { + "path": "string_acoustic_012-031-100.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17225 + }, + { + "path": "flute_acoustic_002-078-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17226 + }, + { + "path": "organ_electronic_113-043-100.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17227 + }, + { + "path": "organ_electronic_007-043-075.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17228 + }, + { + "path": "string_acoustic_014-048-050.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17229 + }, + { + "path": "mallet_acoustic_056-044-025.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17230 + }, + { + "path": "brass_acoustic_059-048-050.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17231 + }, + { + "path": "keyboard_electronic_001-043-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17232 + }, + { + "path": "organ_electronic_028-066-075.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17233 + }, + { + "path": "vocal_synthetic_003-073-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17234 + }, + { + "path": "string_acoustic_056-056-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17235 + }, + { + "path": "mallet_acoustic_047-091-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17236 + }, + { + "path": "keyboard_electronic_002-094-025.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17237 + }, + { + "path": "guitar_acoustic_030-038-075.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17238 + }, + { + "path": "vocal_synthetic_003-036-075.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17239 + }, + { + "path": "keyboard_synthetic_000-062-075.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17240 + }, + { + "path": "bass_electronic_018-060-075.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17241 + }, + { + "path": "keyboard_electronic_002-076-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17242 + }, + { + "path": "bass_synthetic_009-071-050.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17243 + }, + { + "path": "bass_synthetic_134-058-050.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17244 + }, + { + "path": "mallet_acoustic_062-059-127.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17245 + }, + { + "path": "keyboard_electronic_078-074-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17246 + }, + { + "path": "organ_electronic_028-091-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17247 + }, + { + "path": "keyboard_acoustic_004-073-050.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17248 + }, + { + "path": "brass_acoustic_016-085-100.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17249 + }, + { + "path": "keyboard_synthetic_000-049-127.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17250 + }, + { + "path": "bass_synthetic_134-039-127.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17251 + }, + { + "path": "bass_synthetic_134-095-025.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17252 + }, + { + "path": "organ_electronic_057-050-025.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17253 + }, + { + "path": "keyboard_electronic_098-058-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17254 + }, + { + "path": "bass_synthetic_135-040-100.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17255 + }, + { + "path": "brass_acoustic_006-058-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "reed", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17256 + }, + { + "path": "flute_synthetic_000-103-025.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17257 + }, + { + "path": "bass_synthetic_134-108-050.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17258 + }, + { + "path": "keyboard_acoustic_004-091-100.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17259 + }, + { + "path": "keyboard_synthetic_000-047-075.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17260 + }, + { + "path": "bass_synthetic_034-063-025.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17261 + }, + { + "path": "bass_synthetic_033-024-075.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17262 + }, + { + "path": "bass_synthetic_134-105-050.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17263 + }, + { + "path": "reed_acoustic_037-058-050.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17264 + }, + { + "path": "brass_acoustic_015-058-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17265 + }, + { + "path": "brass_acoustic_015-046-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17266 + }, + { + "path": "keyboard_electronic_078-038-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17267 + }, + { + "path": "string_acoustic_056-049-025.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17268 + }, + { + "path": "keyboard_electronic_098-023-100.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17269 + }, + { + "path": "organ_electronic_113-056-075.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17270 + }, + { + "path": "mallet_acoustic_056-083-127.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17271 + }, + { + "path": "organ_electronic_001-040-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17272 + }, + { + "path": "bass_synthetic_034-098-100.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17273 + }, + { + "path": "vocal_synthetic_003-096-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17274 + }, + { + "path": "flute_synthetic_000-065-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17275 + }, + { + "path": "guitar_acoustic_015-082-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17276 + }, + { + "path": "brass_acoustic_059-035-100.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17277 + }, + { + "path": "guitar_acoustic_015-021-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17278 + }, + { + "path": "reed_acoustic_023-073-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17279 + }, + { + "path": "mallet_acoustic_062-050-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17280 + }, + { + "path": "organ_electronic_007-039-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17281 + }, + { + "path": "bass_synthetic_033-030-025.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17282 + }, + { + "path": "keyboard_electronic_078-069-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17283 + }, + { + "path": "organ_electronic_057-052-100.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17284 + }, + { + "path": "bass_synthetic_034-106-075.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17285 + }, + { + "path": "brass_acoustic_059-032-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17286 + }, + { + "path": "bass_synthetic_009-080-100.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17287 + }, + { + "path": "organ_electronic_028-056-050.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17288 + }, + { + "path": "bass_synthetic_068-032-050.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17289 + }, + { + "path": "organ_electronic_007-091-025.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17290 + }, + { + "path": "guitar_electronic_022-059-050.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17291 + }, + { + "path": "keyboard_acoustic_004-045-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17292 + }, + { + "path": "keyboard_acoustic_004-030-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17293 + }, + { + "path": "bass_synthetic_134-073-075.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17294 + }, + { + "path": "reed_acoustic_023-067-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17295 + }, + { + "path": "organ_electronic_057-062-100.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17296 + }, + { + "path": "keyboard_electronic_098-107-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17297 + }, + { + "path": "bass_synthetic_134-075-075.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17298 + }, + { + "path": "organ_electronic_113-047-127.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17299 + }, + { + "path": "bass_synthetic_134-074-025.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17300 + }, + { + "path": "flute_acoustic_002-072-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17301 + }, + { + "path": "keyboard_electronic_069-084-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17302 + }, + { + "path": "guitar_acoustic_021-086-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17303 + }, + { + "path": "keyboard_electronic_078-061-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17304 + }, + { + "path": "reed_acoustic_023-065-050.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17305 + }, + { + "path": "string_acoustic_080-043-127.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17306 + }, + { + "path": "vocal_synthetic_003-049-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17307 + }, + { + "path": "bass_synthetic_098-060-100.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17308 + }, + { + "path": "string_acoustic_014-038-075.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17309 + }, + { + "path": "organ_electronic_007-045-127.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17310 + }, + { + "path": "flute_synthetic_000-070-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17311 + }, + { + "path": "guitar_acoustic_021-090-075.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17312 + }, + { + "path": "brass_acoustic_006-068-025.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17313 + }, + { + "path": "string_acoustic_014-037-100.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17314 + }, + { + "path": "keyboard_electronic_098-043-127.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17315 + }, + { + "path": "keyboard_electronic_003-038-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17316 + }, + { + "path": "keyboard_acoustic_004-050-127.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17317 + }, + { + "path": "keyboard_electronic_078-083-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17318 + }, + { + "path": "keyboard_synthetic_000-072-127.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17319 + }, + { + "path": "string_acoustic_012-034-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17320 + }, + { + "path": "string_acoustic_057-062-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17321 + }, + { + "path": "organ_electronic_007-096-100.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17322 + }, + { + "path": "reed_acoustic_011-044-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17323 + }, + { + "path": "organ_electronic_001-048-025.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17324 + }, + { + "path": "keyboard_electronic_001-104-127.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "flute", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17325 + }, + { + "path": "keyboard_electronic_001-064-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17326 + }, + { + "path": "keyboard_electronic_001-092-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17327 + }, + { + "path": "keyboard_acoustic_004-057-075.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17328 + }, + { + "path": "guitar_acoustic_015-051-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17329 + }, + { + "path": "guitar_electronic_022-036-100.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17330 + }, + { + "path": "keyboard_electronic_002-101-100.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17331 + }, + { + "path": "string_acoustic_056-041-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17332 + }, + { + "path": "mallet_acoustic_062-079-075.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17333 + }, + { + "path": "keyboard_electronic_002-080-050.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17334 + }, + { + "path": "brass_acoustic_046-092-127.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17335 + }, + { + "path": "keyboard_electronic_001-045-050.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17336 + }, + { + "path": "keyboard_electronic_001-059-075.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17337 + }, + { + "path": "bass_synthetic_068-096-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17338 + }, + { + "path": "keyboard_electronic_078-050-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17339 + }, + { + "path": "reed_acoustic_023-056-050.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17340 + }, + { + "path": "bass_synthetic_034-060-127.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17341 + }, + { + "path": "brass_acoustic_046-068-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17342 + }, + { + "path": "brass_acoustic_046-070-127.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17343 + }, + { + "path": "bass_synthetic_134-085-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17344 + }, + { + "path": "bass_synthetic_134-098-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17345 + }, + { + "path": "reed_acoustic_037-041-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17346 + }, + { + "path": "keyboard_electronic_069-061-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17347 + }, + { + "path": "bass_synthetic_034-061-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17348 + }, + { + "path": "guitar_electronic_022-060-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17349 + }, + { + "path": "keyboard_synthetic_000-094-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17350 + }, + { + "path": "keyboard_electronic_002-060-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17351 + }, + { + "path": "bass_synthetic_033-049-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17352 + }, + { + "path": "string_acoustic_080-053-075.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17353 + }, + { + "path": "string_acoustic_080-026-025.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17354 + }, + { + "path": "organ_electronic_057-037-127.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17355 + }, + { + "path": "keyboard_electronic_002-084-025.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17356 + }, + { + "path": "keyboard_electronic_098-099-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17357 + }, + { + "path": "brass_acoustic_006-043-100.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17358 + }, + { + "path": "keyboard_electronic_001-037-075.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17359 + }, + { + "path": "string_acoustic_056-064-025.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17360 + }, + { + "path": "bass_synthetic_033-060-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17361 + }, + { + "path": "brass_acoustic_046-096-075.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17362 + }, + { + "path": "bass_synthetic_068-098-025.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17363 + }, + { + "path": "reed_acoustic_037-073-100.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17364 + }, + { + "path": "bass_synthetic_068-050-100.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17365 + }, + { + "path": "bass_synthetic_009-070-100.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17366 + }, + { + "path": "brass_acoustic_059-052-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17367 + }, + { + "path": "bass_synthetic_009-094-100.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17368 + }, + { + "path": "organ_electronic_007-065-127.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17369 + }, + { + "path": "keyboard_electronic_098-098-127.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17370 + }, + { + "path": "vocal_synthetic_003-021-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17371 + }, + { + "path": "mallet_acoustic_062-038-025.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17372 + }, + { + "path": "organ_electronic_028-023-050.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17373 + }, + { + "path": "brass_acoustic_059-062-025.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17374 + }, + { + "path": "keyboard_electronic_098-079-127.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17375 + }, + { + "path": "flute_acoustic_002-099-050.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17376 + }, + { + "path": "string_acoustic_080-028-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17377 + }, + { + "path": "organ_electronic_113-045-100.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17378 + }, + { + "path": "keyboard_electronic_069-079-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17379 + }, + { + "path": "flute_synthetic_000-074-050.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17380 + }, + { + "path": "keyboard_electronic_001-048-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17381 + }, + { + "path": "brass_acoustic_006-048-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17382 + }, + { + "path": "flute_acoustic_002-103-050.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17383 + }, + { + "path": "bass_synthetic_098-059-050.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17384 + }, + { + "path": "mallet_acoustic_062-045-127.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17385 + }, + { + "path": "brass_acoustic_016-066-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17386 + }, + { + "path": "keyboard_electronic_078-036-100.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17387 + }, + { + "path": "organ_electronic_028-027-127.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17388 + }, + { + "path": "bass_synthetic_098-049-025.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17389 + }, + { + "path": "keyboard_electronic_001-054-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17390 + }, + { + "path": "brass_acoustic_016-060-100.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17391 + }, + { + "path": "keyboard_electronic_098-090-100.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17392 + }, + { + "path": "brass_acoustic_046-088-075.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17393 + }, + { + "path": "brass_acoustic_015-073-075.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17394 + }, + { + "path": "string_acoustic_056-038-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17395 + }, + { + "path": "brass_acoustic_006-045-050.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17396 + }, + { + "path": "bass_synthetic_034-073-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17397 + }, + { + "path": "bass_synthetic_009-079-075.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17398 + }, + { + "path": "guitar_acoustic_010-101-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17399 + }, + { + "path": "brass_acoustic_006-040-100.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17400 + }, + { + "path": "bass_synthetic_009-087-025.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17401 + }, + { + "path": "guitar_acoustic_014-103-075.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17402 + }, + { + "path": "guitar_acoustic_014-085-127.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17403 + }, + { + "path": "brass_acoustic_006-062-050.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17404 + }, + { + "path": "flute_synthetic_000-048-050.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17405 + }, + { + "path": "flute_synthetic_000-087-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17406 + }, + { + "path": "vocal_synthetic_003-083-025.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17407 + }, + { + "path": "mallet_acoustic_062-058-050.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17408 + }, + { + "path": "keyboard_electronic_069-066-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17409 + }, + { + "path": "keyboard_electronic_001-042-025.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17410 + }, + { + "path": "guitar_electronic_028-030-075.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17411 + }, + { + "path": "keyboard_electronic_098-044-025.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17412 + }, + { + "path": "bass_synthetic_068-108-075.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17413 + }, + { + "path": "guitar_acoustic_015-030-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17414 + }, + { + "path": "organ_electronic_028-068-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17415 + }, + { + "path": "bass_synthetic_135-091-127.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17416 + }, + { + "path": "keyboard_electronic_098-097-025.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17417 + }, + { + "path": "mallet_acoustic_056-105-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17418 + }, + { + "path": "brass_acoustic_006-030-127.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17419 + }, + { + "path": "bass_synthetic_135-028-050.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17420 + }, + { + "path": "flute_synthetic_000-068-100.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17421 + }, + { + "path": "vocal_acoustic_000-054-050.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17422 + }, + { + "path": "keyboard_electronic_069-039-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17423 + }, + { + "path": "vocal_synthetic_003-081-100.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17424 + }, + { + "path": "keyboard_electronic_001-045-025.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17425 + }, + { + "path": "reed_acoustic_037-062-050.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17426 + }, + { + "path": "brass_acoustic_006-065-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17427 + }, + { + "path": "keyboard_acoustic_004-096-100.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17428 + }, + { + "path": "mallet_acoustic_062-057-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17429 + }, + { + "path": "bass_synthetic_134-080-100.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17430 + }, + { + "path": "organ_electronic_113-056-127.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17431 + }, + { + "path": "keyboard_electronic_001-046-025.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17432 + }, + { + "path": "keyboard_electronic_002-098-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17433 + }, + { + "path": "brass_acoustic_059-053-025.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17434 + }, + { + "path": "guitar_electronic_028-037-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17435 + }, + { + "path": "keyboard_electronic_069-052-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17436 + }, + { + "path": "bass_synthetic_009-013-127.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17437 + }, + { + "path": "string_acoustic_014-059-025.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17438 + }, + { + "path": "brass_acoustic_059-031-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17439 + }, + { + "path": "keyboard_electronic_002-090-050.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17440 + }, + { + "path": "flute_synthetic_000-082-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17441 + }, + { + "path": "guitar_acoustic_015-102-100.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17442 + }, + { + "path": "bass_synthetic_034-105-127.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17443 + }, + { + "path": "guitar_acoustic_015-080-127.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17444 + }, + { + "path": "organ_electronic_113-034-075.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17445 + }, + { + "path": "brass_acoustic_016-075-127.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "flute", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17446 + }, + { + "path": "bass_synthetic_135-028-100.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17447 + }, + { + "path": "bass_synthetic_134-065-127.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17448 + }, + { + "path": "bass_synthetic_134-085-127.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17449 + }, + { + "path": "organ_electronic_113-051-127.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17450 + }, + { + "path": "organ_electronic_028-103-127.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17451 + }, + { + "path": "bass_electronic_025-022-100.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17452 + }, + { + "path": "bass_synthetic_033-056-050.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17453 + }, + { + "path": "bass_synthetic_033-042-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17454 + }, + { + "path": "keyboard_electronic_001-086-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17455 + }, + { + "path": "string_acoustic_056-040-075.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17456 + }, + { + "path": "keyboard_electronic_002-095-127.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17457 + }, + { + "path": "vocal_synthetic_003-035-127.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17458 + }, + { + "path": "string_acoustic_056-050-127.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17459 + }, + { + "path": "mallet_acoustic_062-045-100.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17460 + }, + { + "path": "bass_synthetic_009-049-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17461 + }, + { + "path": "bass_synthetic_135-024-100.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17462 + }, + { + "path": "vocal_synthetic_003-037-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17463 + }, + { + "path": "keyboard_electronic_003-058-025.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17464 + }, + { + "path": "guitar_acoustic_015-030-075.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17465 + }, + { + "path": "reed_acoustic_037-059-050.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17466 + }, + { + "path": "bass_synthetic_098-027-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17467 + }, + { + "path": "string_acoustic_057-089-025.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17468 + }, + { + "path": "organ_electronic_028-071-127.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17469 + }, + { + "path": "mallet_acoustic_062-067-127.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17470 + }, + { + "path": "keyboard_acoustic_004-046-127.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17471 + }, + { + "path": "guitar_electronic_022-042-075.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17472 + }, + { + "path": "bass_synthetic_135-045-075.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17473 + }, + { + "path": "keyboard_synthetic_000-064-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17474 + }, + { + "path": "guitar_acoustic_021-075-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17475 + }, + { + "path": "vocal_synthetic_003-041-075.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17476 + }, + { + "path": "guitar_acoustic_021-046-127.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17477 + }, + { + "path": "keyboard_electronic_069-048-075.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17478 + }, + { + "path": "string_acoustic_080-051-050.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17479 + }, + { + "path": "keyboard_electronic_069-064-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17480 + }, + { + "path": "keyboard_electronic_001-064-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17481 + }, + { + "path": "guitar_acoustic_010-100-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17482 + }, + { + "path": "mallet_acoustic_062-102-050.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17483 + }, + { + "path": "bass_synthetic_134-100-100.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17484 + }, + { + "path": "bass_synthetic_009-091-050.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17485 + }, + { + "path": "string_acoustic_056-067-050.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17486 + }, + { + "path": "bass_synthetic_134-027-025.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17487 + }, + { + "path": "flute_synthetic_000-095-025.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17488 + }, + { + "path": "organ_electronic_113-057-127.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17489 + }, + { + "path": "organ_electronic_007-041-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17490 + }, + { + "path": "keyboard_electronic_078-045-025.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17491 + }, + { + "path": "keyboard_electronic_069-048-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17492 + }, + { + "path": "string_acoustic_071-026-025.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17493 + }, + { + "path": "bass_synthetic_034-040-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17494 + }, + { + "path": "keyboard_electronic_098-045-075.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17495 + }, + { + "path": "string_acoustic_057-056-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17496 + }, + { + "path": "keyboard_acoustic_004-045-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17497 + }, + { + "path": "mallet_acoustic_047-073-075.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17498 + }, + { + "path": "vocal_synthetic_003-027-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "vocal", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17499 + }, + { + "path": "brass_acoustic_016-063-025.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17500 + }, + { + "path": "bass_synthetic_034-085-075.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "flute", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17501 + }, + { + "path": "bass_synthetic_009-012-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17502 + }, + { + "path": "organ_electronic_028-060-127.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17503 + }, + { + "path": "bass_synthetic_009-067-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17504 + }, + { + "path": "guitar_electronic_028-032-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17505 + }, + { + "path": "organ_electronic_057-030-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17506 + }, + { + "path": "bass_synthetic_068-052-100.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17507 + }, + { + "path": "vocal_synthetic_003-093-100.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "vocal", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17508 + }, + { + "path": "guitar_acoustic_015-081-050.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17509 + }, + { + "path": "bass_synthetic_034-053-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17510 + }, + { + "path": "string_acoustic_056-073-025.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17511 + }, + { + "path": "guitar_acoustic_021-091-127.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17512 + }, + { + "path": "string_acoustic_080-030-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17513 + }, + { + "path": "bass_synthetic_009-096-127.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17514 + }, + { + "path": "keyboard_electronic_003-057-025.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17515 + }, + { + "path": "guitar_acoustic_030-077-100.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17516 + }, + { + "path": "mallet_acoustic_062-042-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17517 + }, + { + "path": "guitar_acoustic_030-064-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17518 + }, + { + "path": "organ_electronic_113-040-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17519 + }, + { + "path": "mallet_acoustic_047-082-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17520 + }, + { + "path": "mallet_acoustic_062-084-100.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17521 + }, + { + "path": "bass_synthetic_068-107-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17522 + }, + { + "path": "vocal_synthetic_003-037-075.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17523 + }, + { + "path": "mallet_acoustic_047-098-127.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17524 + }, + { + "path": "organ_electronic_007-028-025.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17525 + }, + { + "path": "reed_acoustic_023-076-127.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17526 + }, + { + "path": "reed_acoustic_023-095-100.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17527 + }, + { + "path": "bass_synthetic_135-072-050.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17528 + }, + { + "path": "keyboard_electronic_098-040-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17529 + }, + { + "path": "guitar_acoustic_014-105-075.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17530 + }, + { + "path": "brass_acoustic_006-034-100.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17531 + }, + { + "path": "organ_electronic_028-047-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17532 + }, + { + "path": "organ_electronic_057-026-075.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17533 + }, + { + "path": "keyboard_synthetic_000-096-127.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17534 + }, + { + "path": "bass_synthetic_009-009-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17535 + }, + { + "path": "bass_synthetic_009-093-075.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17536 + }, + { + "path": "reed_acoustic_037-044-127.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17537 + }, + { + "path": "bass_synthetic_134-060-025.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17538 + }, + { + "path": "reed_acoustic_023-059-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17539 + }, + { + "path": "vocal_synthetic_003-081-050.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17540 + }, + { + "path": "organ_electronic_113-054-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17541 + }, + { + "path": "guitar_acoustic_015-102-050.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17542 + }, + { + "path": "keyboard_electronic_098-041-100.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17543 + }, + { + "path": "mallet_acoustic_047-093-025.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17544 + }, + { + "path": "mallet_acoustic_047-090-075.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17545 + }, + { + "path": "mallet_acoustic_062-022-025.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17546 + }, + { + "path": "organ_electronic_028-054-127.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17547 + }, + { + "path": "reed_acoustic_023-097-100.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17548 + }, + { + "path": "string_acoustic_071-044-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17549 + }, + { + "path": "string_acoustic_057-069-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17550 + }, + { + "path": "keyboard_electronic_002-033-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17551 + }, + { + "path": "string_acoustic_056-039-075.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17552 + }, + { + "path": "vocal_synthetic_003-043-100.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17553 + }, + { + "path": "keyboard_synthetic_000-044-025.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17554 + }, + { + "path": "bass_synthetic_135-091-025.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17555 + }, + { + "path": "keyboard_electronic_002-026-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17556 + }, + { + "path": "reed_acoustic_023-033-075.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17557 + }, + { + "path": "keyboard_electronic_001-098-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17558 + }, + { + "path": "bass_synthetic_134-050-075.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17559 + }, + { + "path": "mallet_acoustic_062-036-127.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17560 + }, + { + "path": "guitar_acoustic_015-071-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17561 + }, + { + "path": "keyboard_acoustic_004-107-025.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17562 + }, + { + "path": "bass_synthetic_135-108-050.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17563 + }, + { + "path": "string_acoustic_012-055-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17564 + }, + { + "path": "bass_synthetic_134-057-075.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17565 + }, + { + "path": "organ_electronic_028-050-100.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17566 + }, + { + "path": "keyboard_electronic_002-024-050.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17567 + }, + { + "path": "guitar_acoustic_015-105-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17568 + }, + { + "path": "string_acoustic_014-029-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17569 + }, + { + "path": "bass_synthetic_033-050-025.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17570 + }, + { + "path": "reed_acoustic_023-085-075.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17571 + }, + { + "path": "keyboard_synthetic_000-082-127.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17572 + }, + { + "path": "keyboard_electronic_098-024-075.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17573 + }, + { + "path": "organ_electronic_057-060-127.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17574 + }, + { + "path": "bass_synthetic_068-021-100.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17575 + }, + { + "path": "keyboard_acoustic_004-042-127.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17576 + }, + { + "path": "guitar_acoustic_030-085-100.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17577 + }, + { + "path": "keyboard_electronic_002-048-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17578 + }, + { + "path": "guitar_acoustic_015-069-075.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17579 + }, + { + "path": "mallet_acoustic_062-080-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17580 + }, + { + "path": "keyboard_electronic_002-065-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17581 + }, + { + "path": "bass_synthetic_009-052-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "keyboard", + "choice_b": "string", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17582 + }, + { + "path": "string_acoustic_057-059-025.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17583 + }, + { + "path": "organ_electronic_113-045-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17584 + }, + { + "path": "organ_electronic_001-045-100.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17585 + }, + { + "path": "keyboard_electronic_098-046-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "bass", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17586 + }, + { + "path": "keyboard_synthetic_000-021-050.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17587 + }, + { + "path": "string_acoustic_056-044-025.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17588 + }, + { + "path": "guitar_acoustic_015-071-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17589 + }, + { + "path": "guitar_acoustic_030-070-075.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17590 + }, + { + "path": "keyboard_electronic_078-041-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17591 + }, + { + "path": "bass_synthetic_033-023-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17592 + }, + { + "path": "mallet_acoustic_062-033-025.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17593 + }, + { + "path": "bass_synthetic_009-077-075.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17594 + }, + { + "path": "bass_synthetic_068-034-100.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17595 + }, + { + "path": "organ_electronic_028-054-025.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17596 + }, + { + "path": "flute_acoustic_002-090-050.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17597 + }, + { + "path": "keyboard_synthetic_000-079-075.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17598 + }, + { + "path": "keyboard_electronic_002-101-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17599 + }, + { + "path": "bass_synthetic_135-046-127.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17600 + }, + { + "path": "bass_synthetic_034-094-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17601 + }, + { + "path": "string_acoustic_080-052-100.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17602 + }, + { + "path": "mallet_acoustic_062-064-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17603 + }, + { + "path": "keyboard_electronic_002-103-050.wav", + "question": "What's the primary source of sound in this music track?", + "choice_a": "mallet", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17604 + }, + { + "path": "bass_synthetic_098-021-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17605 + }, + { + "path": "brass_acoustic_016-068-127.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17606 + }, + { + "path": "flute_acoustic_002-073-127.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "reed", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17607 + }, + { + "path": "organ_electronic_113-039-050.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17608 + }, + { + "path": "organ_electronic_028-028-100.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17609 + }, + { + "path": "flute_acoustic_002-070-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17610 + }, + { + "path": "keyboard_electronic_098-083-100.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17611 + }, + { + "path": "string_acoustic_057-081-025.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17612 + }, + { + "path": "guitar_acoustic_021-102-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17613 + }, + { + "path": "bass_synthetic_068-045-100.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17614 + }, + { + "path": "brass_acoustic_015-063-025.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17615 + }, + { + "path": "brass_acoustic_015-054-100.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17616 + }, + { + "path": "string_acoustic_056-070-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "guitar", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17617 + }, + { + "path": "keyboard_electronic_002-107-127.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17618 + }, + { + "path": "bass_synthetic_009-029-075.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17619 + }, + { + "path": "bass_synthetic_033-028-050.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17620 + }, + { + "path": "keyboard_electronic_098-055-025.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17621 + }, + { + "path": "guitar_acoustic_030-033-025.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "reed", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17622 + }, + { + "path": "brass_acoustic_016-080-025.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17623 + }, + { + "path": "keyboard_electronic_003-023-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17624 + }, + { + "path": "keyboard_synthetic_000-034-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17625 + }, + { + "path": "keyboard_electronic_001-040-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17626 + }, + { + "path": "keyboard_electronic_002-084-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "string", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17627 + }, + { + "path": "bass_synthetic_134-069-100.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17628 + }, + { + "path": "reed_acoustic_023-046-050.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17629 + }, + { + "path": "bass_synthetic_034-085-025.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17630 + }, + { + "path": "organ_electronic_028-037-025.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17631 + }, + { + "path": "brass_acoustic_046-104-025.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17632 + }, + { + "path": "bass_synthetic_068-046-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17633 + }, + { + "path": "brass_acoustic_006-056-127.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17634 + }, + { + "path": "flute_synthetic_000-101-025.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17635 + }, + { + "path": "organ_electronic_007-070-075.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17636 + }, + { + "path": "keyboard_electronic_098-028-075.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17637 + }, + { + "path": "flute_acoustic_002-081-025.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "flute", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17638 + }, + { + "path": "string_acoustic_080-032-025.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "vocal", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17639 + }, + { + "path": "keyboard_electronic_002-046-100.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17640 + }, + { + "path": "string_acoustic_012-029-127.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17641 + }, + { + "path": "guitar_acoustic_015-081-075.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17642 + }, + { + "path": "guitar_acoustic_015-097-025.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17643 + }, + { + "path": "string_acoustic_071-035-075.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17644 + }, + { + "path": "guitar_acoustic_015-108-127.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17645 + }, + { + "path": "keyboard_electronic_069-032-127.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17646 + }, + { + "path": "keyboard_electronic_001-044-050.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17647 + }, + { + "path": "bass_synthetic_135-085-127.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17648 + }, + { + "path": "guitar_acoustic_021-040-127.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17649 + }, + { + "path": "guitar_acoustic_030-043-025.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17650 + }, + { + "path": "keyboard_acoustic_004-075-025.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17651 + }, + { + "path": "keyboard_synthetic_000-064-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17652 + }, + { + "path": "bass_synthetic_134-034-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17653 + }, + { + "path": "mallet_acoustic_062-073-025.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17654 + }, + { + "path": "keyboard_electronic_003-028-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17655 + }, + { + "path": "bass_synthetic_098-040-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17656 + }, + { + "path": "organ_electronic_028-106-100.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17657 + }, + { + "path": "flute_synthetic_000-100-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17658 + }, + { + "path": "brass_acoustic_006-077-025.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17659 + }, + { + "path": "keyboard_electronic_002-031-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17660 + }, + { + "path": "bass_synthetic_009-083-075.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "vocal", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17661 + }, + { + "path": "guitar_electronic_028-050-100.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17662 + }, + { + "path": "guitar_acoustic_010-107-075.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17663 + }, + { + "path": "flute_synthetic_000-090-050.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "reed", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17664 + }, + { + "path": "brass_acoustic_046-106-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17665 + }, + { + "path": "organ_electronic_007-077-025.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17666 + }, + { + "path": "bass_synthetic_134-044-025.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17667 + }, + { + "path": "vocal_synthetic_003-065-050.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "bass", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17668 + }, + { + "path": "vocal_synthetic_003-021-127.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17669 + }, + { + "path": "keyboard_electronic_001-022-050.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17670 + }, + { + "path": "string_acoustic_014-025-075.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17671 + }, + { + "path": "bass_synthetic_134-080-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17672 + }, + { + "path": "mallet_acoustic_047-065-075.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17673 + }, + { + "path": "reed_acoustic_011-049-025.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17674 + }, + { + "path": "keyboard_electronic_069-046-025.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17675 + }, + { + "path": "bass_synthetic_134-040-025.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17676 + }, + { + "path": "guitar_acoustic_015-094-100.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17677 + }, + { + "path": "bass_synthetic_009-053-075.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17678 + }, + { + "path": "brass_acoustic_006-087-050.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17679 + }, + { + "path": "guitar_electronic_028-035-025.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17680 + }, + { + "path": "bass_synthetic_009-050-100.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17681 + }, + { + "path": "guitar_acoustic_014-106-100.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "vocal", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17682 + }, + { + "path": "mallet_acoustic_062-086-050.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17683 + }, + { + "path": "organ_electronic_007-077-127.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "flute", + "choice_b": "organ", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17684 + }, + { + "path": "brass_acoustic_006-082-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "brass", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17685 + }, + { + "path": "keyboard_electronic_002-099-100.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17686 + }, + { + "path": "reed_acoustic_037-060-075.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17687 + }, + { + "path": "keyboard_synthetic_000-050-075.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17688 + }, + { + "path": "brass_acoustic_006-049-127.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17689 + }, + { + "path": "brass_acoustic_006-073-075.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17690 + }, + { + "path": "guitar_acoustic_010-108-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17691 + }, + { + "path": "keyboard_electronic_098-025-050.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17692 + }, + { + "path": "keyboard_electronic_098-038-127.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17693 + }, + { + "path": "mallet_acoustic_062-078-127.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17694 + }, + { + "path": "brass_acoustic_046-059-075.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17695 + }, + { + "path": "flute_acoustic_002-084-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17696 + }, + { + "path": "guitar_acoustic_014-102-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17697 + }, + { + "path": "organ_electronic_007-011-075.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "mallet", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17698 + }, + { + "path": "keyboard_electronic_078-038-127.wav", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17699 + }, + { + "path": "keyboard_electronic_003-034-050.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17700 + }, + { + "path": "brass_acoustic_016-078-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17701 + }, + { + "path": "bass_electronic_018-061-127.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17702 + }, + { + "path": "bass_synthetic_134-079-050.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "string", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17703 + }, + { + "path": "guitar_acoustic_015-049-025.wav", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17704 + }, + { + "path": "organ_electronic_028-045-025.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "string", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17705 + }, + { + "path": "organ_electronic_007-096-025.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17706 + }, + { + "path": "guitar_acoustic_015-087-100.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17707 + }, + { + "path": "bass_synthetic_068-038-025.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17708 + }, + { + "path": "bass_synthetic_134-044-100.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bass", + "choice_b": "reed", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17709 + }, + { + "path": "organ_electronic_113-034-127.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17710 + }, + { + "path": "organ_electronic_007-092-127.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17711 + }, + { + "path": "brass_acoustic_006-032-127.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17712 + }, + { + "path": "flute_synthetic_000-027-127.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17713 + }, + { + "path": "string_acoustic_057-076-050.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "reed", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17714 + }, + { + "path": "string_acoustic_012-024-025.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17715 + }, + { + "path": "bass_synthetic_034-050-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17716 + }, + { + "path": "keyboard_electronic_001-073-050.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "flute", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17717 + }, + { + "path": "keyboard_electronic_078-040-050.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17718 + }, + { + "path": "keyboard_electronic_002-033-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17719 + }, + { + "path": "vocal_acoustic_000-069-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17720 + }, + { + "path": "keyboard_acoustic_004-030-127.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17721 + }, + { + "path": "brass_acoustic_006-027-050.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17722 + }, + { + "path": "flute_synthetic_000-030-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17723 + }, + { + "path": "bass_synthetic_135-035-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17724 + }, + { + "path": "guitar_electronic_022-052-050.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17725 + }, + { + "path": "bass_synthetic_135-087-075.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17726 + }, + { + "path": "keyboard_acoustic_004-040-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17727 + }, + { + "path": "bass_synthetic_135-030-025.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17728 + }, + { + "path": "guitar_electronic_022-061-127.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17729 + }, + { + "path": "organ_electronic_028-108-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17730 + }, + { + "path": "brass_acoustic_015-058-127.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "keyboard", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17731 + }, + { + "path": "keyboard_acoustic_004-105-025.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17732 + }, + { + "path": "bass_electronic_018-060-127.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17733 + }, + { + "path": "string_acoustic_012-049-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17734 + }, + { + "path": "bass_synthetic_098-050-100.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17735 + }, + { + "path": "reed_acoustic_023-080-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17736 + }, + { + "path": "bass_synthetic_009-032-100.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17737 + }, + { + "path": "string_acoustic_056-063-100.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17738 + }, + { + "path": "organ_electronic_028-072-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17739 + }, + { + "path": "string_acoustic_012-052-050.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "vocal", + "choice_b": "reed", + "choice_c": "string", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17740 + }, + { + "path": "flute_synthetic_000-076-025.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17741 + }, + { + "path": "keyboard_electronic_002-074-127.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17742 + }, + { + "path": "keyboard_synthetic_000-042-075.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "string", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17743 + }, + { + "path": "keyboard_synthetic_000-032-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17744 + }, + { + "path": "bass_synthetic_009-096-075.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17745 + }, + { + "path": "keyboard_acoustic_004-061-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17746 + }, + { + "path": "keyboard_electronic_002-089-127.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17747 + }, + { + "path": "organ_electronic_057-036-050.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17748 + }, + { + "path": "bass_synthetic_068-037-025.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17749 + }, + { + "path": "guitar_acoustic_015-081-100.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17750 + }, + { + "path": "mallet_acoustic_062-061-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "organ", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17751 + }, + { + "path": "guitar_acoustic_015-083-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "mallet", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17752 + }, + { + "path": "guitar_acoustic_015-074-127.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17753 + }, + { + "path": "organ_electronic_007-016-075.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17754 + }, + { + "path": "keyboard_electronic_003-046-127.wav", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17755 + }, + { + "path": "flute_synthetic_000-036-100.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "vocal", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17756 + }, + { + "path": "guitar_acoustic_014-086-050.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17757 + }, + { + "path": "brass_acoustic_006-056-100.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17758 + }, + { + "path": "string_acoustic_057-066-100.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "flute", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17759 + }, + { + "path": "keyboard_acoustic_004-079-075.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17760 + }, + { + "path": "reed_acoustic_037-055-075.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17761 + }, + { + "path": "vocal_synthetic_003-059-075.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "vocal", + "choice_d": "organ", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17762 + }, + { + "path": "keyboard_electronic_002-067-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17763 + }, + { + "path": "bass_synthetic_034-093-075.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "vocal", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17764 + }, + { + "path": "bass_synthetic_068-038-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17765 + }, + { + "path": "organ_electronic_001-047-025.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17766 + }, + { + "path": "bass_synthetic_009-071-075.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17767 + }, + { + "path": "string_acoustic_071-029-025.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "reed", + "choice_d": "bass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17768 + }, + { + "path": "brass_acoustic_046-086-025.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17769 + }, + { + "path": "guitar_acoustic_014-101-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17770 + }, + { + "path": "bass_synthetic_135-062-100.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17771 + }, + { + "path": "keyboard_electronic_003-035-050.wav", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17772 + }, + { + "path": "guitar_acoustic_021-099-075.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17773 + }, + { + "path": "string_acoustic_014-026-127.wav", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "organ", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17774 + }, + { + "path": "brass_acoustic_015-076-050.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "keyboard", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17775 + }, + { + "path": "string_acoustic_014-028-127.wav", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "mallet", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17776 + }, + { + "path": "string_acoustic_056-061-050.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "reed", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17777 + }, + { + "path": "bass_synthetic_135-093-025.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17778 + }, + { + "path": "reed_acoustic_023-043-127.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "keyboard", + "choice_b": "brass", + "choice_c": "vocal", + "choice_d": "reed", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17779 + }, + { + "path": "bass_synthetic_034-100-075.wav", + "question": "What's the central instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17780 + }, + { + "path": "keyboard_electronic_002-041-100.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "organ", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17781 + }, + { + "path": "mallet_acoustic_047-076-127.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17782 + }, + { + "path": "keyboard_synthetic_000-075-100.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "string", + "choice_d": "reed", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17783 + }, + { + "path": "guitar_acoustic_030-058-050.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17784 + }, + { + "path": "string_acoustic_080-032-127.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17785 + }, + { + "path": "guitar_acoustic_021-103-127.wav", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17786 + }, + { + "path": "bass_synthetic_009-041-050.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17787 + }, + { + "path": "keyboard_acoustic_004-022-025.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "mallet", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17788 + }, + { + "path": "bass_synthetic_034-022-050.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17789 + }, + { + "path": "guitar_acoustic_014-099-025.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17790 + }, + { + "path": "bass_synthetic_068-084-127.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17791 + }, + { + "path": "reed_acoustic_023-073-025.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17792 + }, + { + "path": "bass_synthetic_068-046-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17793 + }, + { + "path": "reed_acoustic_023-086-075.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "mallet", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17794 + }, + { + "path": "guitar_acoustic_014-088-025.wav", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17795 + }, + { + "path": "keyboard_electronic_001-024-075.wav", + "question": "detect the main instrument that defines this track.", + "choice_a": "brass", + "choice_b": "string", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17796 + }, + { + "path": "bass_synthetic_068-075-127.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "mallet", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17797 + }, + { + "path": "brass_acoustic_006-036-100.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "mallet", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17798 + }, + { + "path": "organ_electronic_007-081-025.wav", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "brass", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17799 + }, + { + "path": "keyboard_electronic_001-063-050.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17800 + }, + { + "path": "keyboard_electronic_002-096-025.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17801 + }, + { + "path": "bass_synthetic_068-022-100.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "reed", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17802 + }, + { + "path": "brass_acoustic_016-067-075.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "brass", + "choice_b": "reed", + "choice_c": "vocal", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17803 + }, + { + "path": "keyboard_electronic_069-060-075.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17804 + }, + { + "path": "brass_acoustic_046-089-050.wav", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17805 + }, + { + "path": "brass_acoustic_046-095-050.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17806 + }, + { + "path": "bass_synthetic_134-075-025.wav", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17807 + }, + { + "path": "bass_synthetic_034-074-050.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17808 + }, + { + "path": "bass_synthetic_034-080-100.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17809 + }, + { + "path": "organ_electronic_001-061-100.wav", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "keyboard", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17810 + }, + { + "path": "flute_synthetic_000-088-025.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17811 + }, + { + "path": "string_acoustic_057-088-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "string", + "choice_d": "brass", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17812 + }, + { + "path": "keyboard_synthetic_000-103-025.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17813 + }, + { + "path": "reed_acoustic_023-058-075.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "reed", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "vocal", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17814 + }, + { + "path": "guitar_acoustic_021-077-075.wav", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "string", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17815 + }, + { + "path": "bass_synthetic_009-014-050.wav", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17816 + }, + { + "path": "guitar_acoustic_030-073-050.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17817 + }, + { + "path": "reed_acoustic_023-097-025.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "mallet", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17818 + }, + { + "path": "brass_acoustic_006-043-127.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "bass", + "choice_b": "keyboard", + "choice_c": "reed", + "choice_d": "brass", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17819 + }, + { + "path": "keyboard_electronic_002-030-075.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "string", + "choice_b": "keyboard", + "choice_c": "mallet", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17820 + }, + { + "path": "brass_acoustic_015-066-100.wav", + "question": "What's the leading instrument featured in this track?", + "choice_a": "keyboard", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17821 + }, + { + "path": "bass_synthetic_098-032-100.wav", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17822 + }, + { + "path": "organ_electronic_001-058-075.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "mallet", + "choice_d": "vocal", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17823 + }, + { + "path": "keyboard_electronic_001-100-050.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "reed", + "choice_b": "keyboard", + "choice_c": "vocal", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17824 + }, + { + "path": "organ_electronic_007-038-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "bass", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17825 + }, + { + "path": "reed_acoustic_023-097-127.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "flute", + "choice_b": "reed", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17826 + }, + { + "path": "bass_synthetic_033-061-127.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "reed", + "choice_b": "brass", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17827 + }, + { + "path": "brass_acoustic_046-091-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17828 + }, + { + "path": "bass_synthetic_068-036-025.wav", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "mallet", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17829 + }, + { + "path": "keyboard_electronic_069-077-127.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "reed", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17830 + }, + { + "path": "keyboard_acoustic_004-040-025.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "mallet", + "choice_b": "string", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17831 + }, + { + "path": "flute_synthetic_000-029-127.wav", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "reed", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17832 + }, + { + "path": "brass_acoustic_059-063-127.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "brass", + "choice_b": "bass", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17833 + }, + { + "path": "bass_synthetic_135-040-025.wav", + "question": "What instrument is the base of this musical piece?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17834 + }, + { + "path": "vocal_synthetic_003-102-075.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "string", + "choice_b": "vocal", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17835 + }, + { + "path": "keyboard_electronic_069-062-050.wav", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "mallet", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17836 + }, + { + "path": "keyboard_synthetic_000-099-050.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17837 + }, + { + "path": "keyboard_electronic_098-056-025.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "brass", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17838 + }, + { + "path": "mallet_acoustic_047-104-127.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "mallet", + "choice_c": "flute", + "choice_d": "string", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17839 + }, + { + "path": "keyboard_synthetic_000-044-100.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "guitar", + "choice_b": "vocal", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17840 + }, + { + "path": "mallet_acoustic_062-058-100.wav", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "brass", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17841 + }, + { + "path": "keyboard_electronic_003-029-075.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "guitar", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17842 + }, + { + "path": "mallet_acoustic_047-089-050.wav", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17843 + }, + { + "path": "bass_synthetic_134-055-025.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "mallet", + "choice_b": "reed", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17844 + }, + { + "path": "brass_acoustic_046-069-075.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17845 + }, + { + "path": "keyboard_synthetic_000-022-025.wav", + "question": "What instrument is the heart of this musical track?", + "choice_a": "string", + "choice_b": "mallet", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17846 + }, + { + "path": "reed_acoustic_023-057-127.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "reed", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "flute", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17847 + }, + { + "path": "bass_synthetic_034-064-100.wav", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "mallet", + "choice_b": "vocal", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17848 + }, + { + "path": "string_acoustic_057-089-127.wav", + "question": "What's the key instrument played in this music piece?", + "choice_a": "vocal", + "choice_b": "string", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17849 + }, + { + "path": "guitar_acoustic_014-098-127.wav", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "organ", + "choice_b": "mallet", + "choice_c": "reed", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17850 + }, + { + "path": "reed_acoustic_023-039-050.wav", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "guitar", + "choice_b": "reed", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "reed", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17851 + }, + { + "path": "guitar_acoustic_021-077-100.wav", + "question": "What instrument is the essence of this musical track?", + "choice_a": "organ", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17852 + }, + { + "path": "string_acoustic_057-071-075.wav", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "vocal", + "choice_b": "brass", + "choice_c": "organ", + "choice_d": "string", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17853 + }, + { + "path": "vocal_synthetic_003-091-025.wav", + "question": "Which instrument takes center stage in this music?", + "choice_a": "vocal", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "vocal", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17854 + }, + { + "path": "organ_electronic_057-054-025.wav", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17855 + }, + { + "path": "string_acoustic_057-058-075.wav", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "string", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "vocal", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17856 + }, + { + "path": "keyboard_electronic_069-070-100.wav", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "bass", + "choice_b": "string", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17857 + }, + { + "path": "string_acoustic_080-028-025.wav", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "string", + "choice_b": "reed", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17858 + }, + { + "path": "string_acoustic_057-085-025.wav", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "string", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "mallet", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17859 + }, + { + "path": "flute_synthetic_000-040-100.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "bass", + "choice_b": "organ", + "choice_c": "flute", + "choice_d": "mallet", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17860 + }, + { + "path": "organ_electronic_028-068-050.wav", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17861 + }, + { + "path": "organ_electronic_007-037-025.wav", + "question": "Which instrument is the core of this composition?", + "choice_a": "organ", + "choice_b": "string", + "choice_c": "bass", + "choice_d": "reed", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17862 + }, + { + "path": "keyboard_acoustic_004-056-050.wav", + "question": "discern the principal instrument in this tune.", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17863 + }, + { + "path": "bass_electronic_027-023-025.wav", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "bass", + "choice_b": "vocal", + "choice_c": "flute", + "choice_d": "brass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17864 + }, + { + "path": "keyboard_electronic_098-044-075.wav", + "question": "What instrument is most prominent in this piece?", + "choice_a": "reed", + "choice_b": "mallet", + "choice_c": "keyboard", + "choice_d": "string", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17865 + }, + { + "path": "mallet_acoustic_062-049-075.wav", + "question": "What instrument carries the main theme in this track?", + "choice_a": "mallet", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "reed", + "answer_gt": "mallet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17866 + }, + { + "path": "guitar_acoustic_015-041-050.wav", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "mallet", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17867 + }, + { + "path": "keyboard_electronic_002-093-075.wav", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "mallet", + "choice_b": "brass", + "choice_c": "keyboard", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17868 + }, + { + "path": "keyboard_electronic_003-028-100.wav", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "keyboard", + "choice_b": "mallet", + "choice_c": "string", + "choice_d": "bass", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17869 + }, + { + "path": "keyboard_electronic_002-105-075.wav", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "bass", + "choice_d": "vocal", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17870 + }, + { + "path": "string_acoustic_071-038-127.wav", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "string", + "choice_b": "bass", + "choice_c": "vocal", + "choice_d": "keyboard", + "answer_gt": "string", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "nsynth", + "uniq_id": 17871 + }, + { + "path": "1329502.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "trumpet", + "choice_b": "horn", + "choice_c": "pipeorgan", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17872 + }, + { + "path": "1361475.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "drums", + "choice_b": "voice", + "choice_c": "electricguitar", + "choice_d": "bass", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17873 + }, + { + "path": "1418661.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "piano", + "choice_b": "harmonica", + "choice_c": "acousticguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17874 + }, + { + "path": "978932.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "synthesizer", + "choice_b": "saxophone", + "choice_c": "drums", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17875 + }, + { + "path": "951886.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "flute", + "choice_b": "harp", + "choice_c": "organ", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17876 + }, + { + "path": "1247040.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "violin", + "choice_b": "percussion", + "choice_c": "beat", + "choice_d": "pad", + "answer_gt": "pad", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17877 + }, + { + "path": "1151403.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "strings", + "choice_b": "piano", + "choice_c": "synthesizer", + "choice_d": "flute", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17878 + }, + { + "path": "477306.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "electricpiano", + "choice_b": "orchestra", + "choice_c": "sampler", + "choice_d": "violin", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17879 + }, + { + "path": "25225.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "piano", + "choice_b": "electricguitar", + "choice_c": "trumpet", + "choice_d": "beat", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17880 + }, + { + "path": "390004.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "cello", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17881 + }, + { + "path": "875374.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "synthesizer", + "choice_b": "sampler", + "choice_c": "classicalguitar", + "choice_d": "drummachine", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17882 + }, + { + "path": "477308.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "drums", + "choice_b": "rhodes", + "choice_c": "orchestra", + "choice_d": "horn", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17883 + }, + { + "path": "1150140.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "cello", + "choice_b": "strings", + "choice_c": "synthesizer", + "choice_d": "computer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17884 + }, + { + "path": "1248613.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "computer", + "choice_b": "drums", + "choice_c": "keyboard", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17885 + }, + { + "path": "1163965.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "organ", + "choice_b": "trumpet", + "choice_c": "drums", + "choice_d": "acousticguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17886 + }, + { + "path": "1054368.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "trumpet", + "choice_b": "piano", + "choice_c": "electricguitar", + "choice_d": "beat", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17887 + }, + { + "path": "1205197.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "violin", + "choice_b": "orchestra", + "choice_c": "computer", + "choice_d": "bell", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17888 + }, + { + "path": "958857.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "beat", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17889 + }, + { + "path": "965753.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "harmonica", + "choice_b": "guitar", + "choice_c": "clarinet", + "choice_d": "violin", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17890 + }, + { + "path": "984997.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "sampler", + "choice_b": "drums", + "choice_c": "piano", + "choice_d": "drummachine", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17891 + }, + { + "path": "1077725.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "drums", + "choice_b": "harp", + "choice_c": "piano", + "choice_d": "sampler", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17892 + }, + { + "path": "1416196.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "drummachine", + "choice_b": "synthesizer", + "choice_c": "flute", + "choice_d": "strings", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17893 + }, + { + "path": "1337944.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "bell", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17894 + }, + { + "path": "858548.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "flute", + "choice_b": "synthesizer", + "choice_c": "piano", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17895 + }, + { + "path": "1347897.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "strings", + "choice_b": "piano", + "choice_c": "orchestra", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17896 + }, + { + "path": "1185651.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "synthesizer", + "choice_b": "saxophone", + "choice_c": "piano", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17897 + }, + { + "path": "1419758.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "computer", + "choice_b": "saxophone", + "choice_c": "accordion", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17898 + }, + { + "path": "155325.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "flute", + "choice_b": "trumpet", + "choice_c": "guitar", + "choice_d": "pad", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17899 + }, + { + "path": "1168541.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "electricpiano", + "choice_b": "rhodes", + "choice_c": "synthesizer", + "choice_d": "horn", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17900 + }, + { + "path": "1293738.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "beat", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17901 + }, + { + "path": "257178.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "drums", + "choice_b": "percussion", + "choice_c": "piano", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17902 + }, + { + "path": "1203656.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "voice", + "choice_b": "sampler", + "choice_c": "beat", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17903 + }, + { + "path": "967573.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "strings", + "choice_b": "flute", + "choice_c": "piano", + "choice_d": "classicalguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17904 + }, + { + "path": "1418657.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "guitar", + "choice_b": "accordion", + "choice_c": "keyboard", + "choice_d": "sampler", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17905 + }, + { + "path": "1268991.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "drummachine", + "choice_b": "guitar", + "choice_c": "accordion", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17906 + }, + { + "path": "103299.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "drums", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "orchestra", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17907 + }, + { + "path": "1031109.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "synthesizer", + "choice_c": "piano", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17908 + }, + { + "path": "1164342.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "strings", + "choice_b": "bass", + "choice_c": "clarinet", + "choice_d": "drummachine", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17909 + }, + { + "path": "944141.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "violin", + "choice_b": "cello", + "choice_c": "piano", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17910 + }, + { + "path": "1083537.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "classicalguitar", + "choice_b": "saxophone", + "choice_c": "keyboard", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17911 + }, + { + "path": "1121532.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "horn", + "choice_b": "acousticguitar", + "choice_c": "classicalguitar", + "choice_d": "brass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17912 + }, + { + "path": "859975.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "computer", + "choice_b": "trumpet", + "choice_c": "flute", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17913 + }, + { + "path": "305537.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "beat", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17914 + }, + { + "path": "1159864.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "flute", + "choice_b": "pad", + "choice_c": "clarinet", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17915 + }, + { + "path": "1276174.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "organ", + "choice_b": "electricguitar", + "choice_c": "clarinet", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17916 + }, + { + "path": "1264830.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "harp", + "choice_b": "harmonica", + "choice_c": "trumpet", + "choice_d": "bell", + "answer_gt": "bell", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17917 + }, + { + "path": "251397.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "synthesizer", + "choice_b": "voice", + "choice_c": "piano", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17918 + }, + { + "path": "1267743.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "piano", + "choice_b": "brass", + "choice_c": "acousticguitar", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17919 + }, + { + "path": "702500.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "keyboard", + "choice_b": "orchestra", + "choice_c": "acousticguitar", + "choice_d": "violin", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17920 + }, + { + "path": "1414813.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "trumpet", + "choice_b": "guitar", + "choice_c": "classicalguitar", + "choice_d": "orchestra", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17921 + }, + { + "path": "1271781.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "bass", + "choice_b": "violin", + "choice_c": "brass", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17922 + }, + { + "path": "30221.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "drums", + "choice_b": "drummachine", + "choice_c": "guitar", + "choice_d": "rhodes", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17923 + }, + { + "path": "1349532.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bell", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17924 + }, + { + "path": "1048230.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "flute", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17925 + }, + { + "path": "1374325.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "harmonica", + "choice_b": "cello", + "choice_c": "electricguitar", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17926 + }, + { + "path": "1154279.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "beat", + "choice_b": "cello", + "choice_c": "saxophone", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17927 + }, + { + "path": "80789.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "electricpiano", + "choice_b": "piano", + "choice_c": "drummachine", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17928 + }, + { + "path": "881026.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "trumpet", + "choice_b": "harmonica", + "choice_c": "piano", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17929 + }, + { + "path": "1354699.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "pad", + "choice_b": "percussion", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17930 + }, + { + "path": "1099456.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "organ", + "choice_b": "synthesizer", + "choice_c": "guitar", + "choice_d": "computer", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17931 + }, + { + "path": "6607.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "accordion", + "choice_b": "violin", + "choice_c": "sampler", + "choice_d": "pipeorgan", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17932 + }, + { + "path": "1244450.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "drums", + "choice_b": "guitar", + "choice_c": "rhodes", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17933 + }, + { + "path": "1348602.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "pad", + "choice_b": "horn", + "choice_c": "drums", + "choice_d": "orchestra", + "answer_gt": "pad", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17934 + }, + { + "path": "669142.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "computer", + "choice_b": "beat", + "choice_c": "acousticguitar", + "choice_d": "drummachine", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17935 + }, + { + "path": "1109633.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "synthesizer", + "choice_b": "voice", + "choice_c": "drums", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17936 + }, + { + "path": "1346876.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "drummachine", + "choice_b": "piano", + "choice_c": "voice", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17937 + }, + { + "path": "1174154.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "clarinet", + "choice_b": "keyboard", + "choice_c": "cello", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17938 + }, + { + "path": "1266727.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "acousticguitar", + "choice_b": "synthesizer", + "choice_c": "clarinet", + "choice_d": "percussion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17939 + }, + { + "path": "1168553.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "piano", + "choice_b": "flute", + "choice_c": "electricpiano", + "choice_d": "classicalguitar", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17940 + }, + { + "path": "1347950.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "percussion", + "choice_b": "brass", + "choice_c": "synthesizer", + "choice_d": "bass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17941 + }, + { + "path": "1365997.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "drums", + "choice_b": "guitar", + "choice_c": "piano", + "choice_d": "acousticguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17942 + }, + { + "path": "736790.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "saxophone", + "choice_b": "organ", + "choice_c": "cello", + "choice_d": "strings", + "answer_gt": "cello", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17943 + }, + { + "path": "1371951.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "orchestra", + "choice_b": "horn", + "choice_c": "synthesizer", + "choice_d": "brass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17944 + }, + { + "path": "1348665.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "beat", + "choice_b": "electricpiano", + "choice_c": "sampler", + "choice_d": "rhodes", + "answer_gt": "rhodes", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17945 + }, + { + "path": "944148.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "computer", + "choice_b": "horn", + "choice_c": "organ", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17946 + }, + { + "path": "921507.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "electricguitar", + "choice_b": "acousticguitar", + "choice_c": "horn", + "choice_d": "saxophone", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17947 + }, + { + "path": "1168555.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "pad", + "choice_b": "voice", + "choice_c": "electricpiano", + "choice_d": "bell", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17948 + }, + { + "path": "979863.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "brass", + "choice_b": "synthesizer", + "choice_c": "harp", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17949 + }, + { + "path": "1163778.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "harp", + "choice_b": "bell", + "choice_c": "percussion", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17950 + }, + { + "path": "88563.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "piano", + "choice_b": "pipeorgan", + "choice_c": "horn", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17951 + }, + { + "path": "562660.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "piano", + "choice_c": "percussion", + "choice_d": "bell", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17952 + }, + { + "path": "914765.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "bell", + "choice_c": "harp", + "choice_d": "synthesizer", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17953 + }, + { + "path": "1251473.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "pad", + "choice_b": "rhodes", + "choice_c": "strings", + "choice_d": "acousticguitar", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17954 + }, + { + "path": "809891.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "classicalguitar", + "choice_b": "drummachine", + "choice_c": "horn", + "choice_d": "electricguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17955 + }, + { + "path": "989717.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "pad", + "choice_b": "piano", + "choice_c": "classicalguitar", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17956 + }, + { + "path": "979845.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "saxophone", + "choice_b": "computer", + "choice_c": "strings", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17957 + }, + { + "path": "1234546.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "saxophone", + "choice_c": "harp", + "choice_d": "orchestra", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17958 + }, + { + "path": "414046.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "cello", + "choice_b": "voice", + "choice_c": "piano", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17959 + }, + { + "path": "1342581.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "voice", + "choice_b": "rhodes", + "choice_c": "guitar", + "choice_d": "strings", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17960 + }, + { + "path": "330211.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "rhodes", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17961 + }, + { + "path": "658065.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "synthesizer", + "choice_c": "electricpiano", + "choice_d": "pad", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17962 + }, + { + "path": "866709.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "sampler", + "choice_b": "horn", + "choice_c": "orchestra", + "choice_d": "trumpet", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17963 + }, + { + "path": "1226355.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "synthesizer", + "choice_b": "clarinet", + "choice_c": "bass", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17964 + }, + { + "path": "1225296.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "acousticguitar", + "choice_b": "bell", + "choice_c": "rhodes", + "choice_d": "strings", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17965 + }, + { + "path": "1192553.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "guitar", + "choice_b": "voice", + "choice_c": "bell", + "choice_d": "strings", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17966 + }, + { + "path": "1280439.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "computer", + "choice_b": "synthesizer", + "choice_c": "classicalguitar", + "choice_d": "pipeorgan", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17967 + }, + { + "path": "1313338.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "classicalguitar", + "choice_b": "beat", + "choice_c": "harp", + "choice_d": "sampler", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17968 + }, + { + "path": "944144.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "trumpet", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17969 + }, + { + "path": "82647.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "brass", + "choice_b": "synthesizer", + "choice_c": "percussion", + "choice_d": "acousticguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17970 + }, + { + "path": "25238.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "bell", + "choice_b": "synthesizer", + "choice_c": "piano", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17971 + }, + { + "path": "945026.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "voice", + "choice_b": "orchestra", + "choice_c": "bell", + "choice_d": "pad", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17972 + }, + { + "path": "1359937.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "bass", + "choice_b": "trumpet", + "choice_c": "drummachine", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17973 + }, + { + "path": "1356520.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "percussion", + "choice_b": "voice", + "choice_c": "clarinet", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17974 + }, + { + "path": "244958.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "harmonica", + "choice_b": "piano", + "choice_c": "clarinet", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17975 + }, + { + "path": "1286514.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "clarinet", + "choice_b": "piano", + "choice_c": "accordion", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17976 + }, + { + "path": "1301829.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "computer", + "choice_b": "electricguitar", + "choice_c": "harmonica", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17977 + }, + { + "path": "184243.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "piano", + "choice_b": "voice", + "choice_c": "horn", + "choice_d": "harmonica", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17978 + }, + { + "path": "1053511.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "orchestra", + "choice_b": "classicalguitar", + "choice_c": "synthesizer", + "choice_d": "sampler", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17979 + }, + { + "path": "969345.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "classicalguitar", + "choice_c": "beat", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17980 + }, + { + "path": "1298571.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "pipeorgan", + "choice_b": "synthesizer", + "choice_c": "horn", + "choice_d": "accordion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17981 + }, + { + "path": "1119164.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "trumpet", + "choice_b": "guitar", + "choice_c": "classicalguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17982 + }, + { + "path": "944145.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "drums", + "choice_b": "bell", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17983 + }, + { + "path": "1347955.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "acousticguitar", + "choice_b": "harp", + "choice_c": "synthesizer", + "choice_d": "trumpet", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17984 + }, + { + "path": "782164.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "cello", + "choice_b": "percussion", + "choice_c": "harp", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17985 + }, + { + "path": "979866.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "synthesizer", + "choice_b": "beat", + "choice_c": "drums", + "choice_d": "keyboard", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17986 + }, + { + "path": "1415941.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "piano", + "choice_c": "bell", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17987 + }, + { + "path": "1029512.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "trumpet", + "choice_b": "pad", + "choice_c": "voice", + "choice_d": "drums", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17988 + }, + { + "path": "1086650.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "strings", + "choice_b": "saxophone", + "choice_c": "violin", + "choice_d": "electricpiano", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17989 + }, + { + "path": "1225293.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "bell", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17990 + }, + { + "path": "859974.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "trumpet", + "choice_b": "harp", + "choice_c": "synthesizer", + "choice_d": "saxophone", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17991 + }, + { + "path": "1286494.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "strings", + "choice_b": "keyboard", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17992 + }, + { + "path": "35385.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "cello", + "choice_b": "piano", + "choice_c": "flute", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17993 + }, + { + "path": "1131682.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "computer", + "choice_b": "harmonica", + "choice_c": "percussion", + "choice_d": "keyboard", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17994 + }, + { + "path": "1116268.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "classicalguitar", + "choice_b": "drums", + "choice_c": "pad", + "choice_d": "rhodes", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17995 + }, + { + "path": "287785.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "drummachine", + "choice_b": "synthesizer", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17996 + }, + { + "path": "1183237.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "violin", + "choice_b": "flute", + "choice_c": "rhodes", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17997 + }, + { + "path": "1418659.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "guitar", + "choice_b": "acousticguitar", + "choice_c": "strings", + "choice_d": "orchestra", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17998 + }, + { + "path": "30227.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "computer", + "choice_b": "guitar", + "choice_c": "bell", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 17999 + }, + { + "path": "1250149.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "sampler", + "choice_b": "electricguitar", + "choice_c": "bell", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18000 + }, + { + "path": "1395656.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "keyboard", + "choice_b": "harmonica", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18001 + }, + { + "path": "1289386.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "electricguitar", + "choice_b": "beat", + "choice_c": "saxophone", + "choice_d": "drums", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18002 + }, + { + "path": "74750.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "horn", + "choice_b": "synthesizer", + "choice_c": "acousticguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18003 + }, + { + "path": "1349531.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "pad", + "choice_b": "piano", + "choice_c": "acousticguitar", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18004 + }, + { + "path": "147597.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "harmonica", + "choice_b": "flute", + "choice_c": "pad", + "choice_d": "computer", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18005 + }, + { + "path": "80755.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "beat", + "choice_b": "piano", + "choice_c": "voice", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18006 + }, + { + "path": "1174089.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "electricguitar", + "choice_b": "saxophone", + "choice_c": "harmonica", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18007 + }, + { + "path": "921510.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "computer", + "choice_b": "pad", + "choice_c": "harp", + "choice_d": "saxophone", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18008 + }, + { + "path": "1347939.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "bell", + "choice_b": "electricpiano", + "choice_c": "clarinet", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18009 + }, + { + "path": "1210606.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "piano", + "choice_b": "cello", + "choice_c": "electricguitar", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18010 + }, + { + "path": "903589.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "piano", + "choice_b": "orchestra", + "choice_c": "acousticguitar", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18011 + }, + { + "path": "1393141.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "rhodes", + "choice_b": "saxophone", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18012 + }, + { + "path": "1284781.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "horn", + "choice_b": "guitar", + "choice_c": "electricpiano", + "choice_d": "clarinet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18013 + }, + { + "path": "1383742.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "bass", + "choice_b": "saxophone", + "choice_c": "voice", + "choice_d": "clarinet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18014 + }, + { + "path": "858536.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bell", + "choice_b": "piano", + "choice_c": "bass", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18015 + }, + { + "path": "1376447.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "accordion", + "choice_b": "strings", + "choice_c": "drums", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18016 + }, + { + "path": "1103548.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "drums", + "choice_b": "violin", + "choice_c": "sampler", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18017 + }, + { + "path": "398186.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "electricguitar", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18018 + }, + { + "path": "1359936.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "synthesizer", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18019 + }, + { + "path": "287979.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "synthesizer", + "choice_b": "harmonica", + "choice_c": "organ", + "choice_d": "pipeorgan", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18020 + }, + { + "path": "1219404.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "percussion", + "choice_b": "synthesizer", + "choice_c": "brass", + "choice_d": "flute", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18021 + }, + { + "path": "1416189.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "orchestra", + "choice_b": "bell", + "choice_c": "harp", + "choice_d": "organ", + "answer_gt": "organ", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18022 + }, + { + "path": "709628.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "rhodes", + "choice_b": "piano", + "choice_c": "keyboard", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18023 + }, + { + "path": "903595.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "violin", + "choice_b": "piano", + "choice_c": "trumpet", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18024 + }, + { + "path": "840600.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "drummachine", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "sampler", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18025 + }, + { + "path": "542788.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "sampler", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18026 + }, + { + "path": "847262.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "violin", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "beat", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18027 + }, + { + "path": "1157349.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "trumpet", + "choice_b": "horn", + "choice_c": "sampler", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18028 + }, + { + "path": "384150.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "trumpet", + "choice_b": "voice", + "choice_c": "beat", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18029 + }, + { + "path": "1053506.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "sampler", + "choice_b": "synthesizer", + "choice_c": "bass", + "choice_d": "saxophone", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18030 + }, + { + "path": "589414.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "voice", + "choice_b": "piano", + "choice_c": "electricguitar", + "choice_d": "classicalguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18031 + }, + { + "path": "1133071.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "percussion", + "choice_b": "pad", + "choice_c": "clarinet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18032 + }, + { + "path": "1347946.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "rhodes", + "choice_b": "accordion", + "choice_c": "electricguitar", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18033 + }, + { + "path": "1143977.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "sampler", + "choice_b": "acousticguitar", + "choice_c": "orchestra", + "choice_d": "flute", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18034 + }, + { + "path": "916396.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "pad", + "choice_b": "horn", + "choice_c": "electricpiano", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18035 + }, + { + "path": "1366000.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "drummachine", + "choice_b": "pipeorgan", + "choice_c": "guitar", + "choice_d": "synthesizer", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18036 + }, + { + "path": "989716.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "horn", + "choice_b": "piano", + "choice_c": "trumpet", + "choice_d": "organ", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18037 + }, + { + "path": "759886.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "acousticguitar", + "choice_b": "guitar", + "choice_c": "harp", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18038 + }, + { + "path": "888327.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "violin", + "choice_b": "electricpiano", + "choice_c": "cello", + "choice_d": "sampler", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18039 + }, + { + "path": "1197390.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "drummachine", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18040 + }, + { + "path": "1173332.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "pad", + "choice_b": "acousticguitar", + "choice_c": "saxophone", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18041 + }, + { + "path": "1171762.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "keyboard", + "choice_b": "harmonica", + "choice_c": "drummachine", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18042 + }, + { + "path": "1087515.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "horn", + "choice_b": "piano", + "choice_c": "drums", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18043 + }, + { + "path": "1402649.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "strings", + "choice_b": "trumpet", + "choice_c": "accordion", + "choice_d": "violin", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18044 + }, + { + "path": "1117567.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "classicalguitar", + "choice_b": "flute", + "choice_c": "accordion", + "choice_d": "trumpet", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18045 + }, + { + "path": "1280423.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "computer", + "choice_b": "synthesizer", + "choice_c": "percussion", + "choice_d": "classicalguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18046 + }, + { + "path": "951321.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "drummachine", + "choice_b": "keyboard", + "choice_c": "saxophone", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18047 + }, + { + "path": "80777.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "bass", + "choice_c": "orchestra", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18048 + }, + { + "path": "1053510.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "harp", + "choice_b": "saxophone", + "choice_c": "synthesizer", + "choice_d": "flute", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18049 + }, + { + "path": "284169.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "clarinet", + "choice_b": "brass", + "choice_c": "electricguitar", + "choice_d": "violin", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18050 + }, + { + "path": "1056630.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "voice", + "choice_b": "brass", + "choice_c": "piano", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18051 + }, + { + "path": "1397347.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "percussion", + "choice_b": "piano", + "choice_c": "bell", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18052 + }, + { + "path": "6606.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "accordion", + "choice_b": "strings", + "choice_c": "acousticguitar", + "choice_d": "harp", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18053 + }, + { + "path": "1374324.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "piano", + "choice_b": "drums", + "choice_c": "trumpet", + "choice_d": "cello", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18054 + }, + { + "path": "475677.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "piano", + "choice_b": "strings", + "choice_c": "organ", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18055 + }, + { + "path": "1204665.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "guitar", + "choice_b": "accordion", + "choice_c": "bell", + "choice_d": "strings", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18056 + }, + { + "path": "911615.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "electricpiano", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "accordion", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18057 + }, + { + "path": "80822.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "clarinet", + "choice_b": "electricpiano", + "choice_c": "saxophone", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18058 + }, + { + "path": "2967.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "acousticguitar", + "choice_b": "guitar", + "choice_c": "voice", + "choice_d": "strings", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18059 + }, + { + "path": "1083510.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "classicalguitar", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "keyboard", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18060 + }, + { + "path": "1209204.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "piano", + "choice_b": "harp", + "choice_c": "beat", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18061 + }, + { + "path": "1210875.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "bell", + "choice_c": "strings", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18062 + }, + { + "path": "979859.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "drums", + "choice_b": "electricpiano", + "choice_c": "piano", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18063 + }, + { + "path": "1286518.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "pipeorgan", + "choice_b": "bell", + "choice_c": "piano", + "choice_d": "organ", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18064 + }, + { + "path": "1313075.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "piano", + "choice_b": "cello", + "choice_c": "drums", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18065 + }, + { + "path": "834109.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "pad", + "choice_b": "voice", + "choice_c": "beat", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18066 + }, + { + "path": "1173259.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "saxophone", + "choice_b": "flute", + "choice_c": "electricpiano", + "choice_d": "computer", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18067 + }, + { + "path": "38282.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "cello", + "choice_b": "horn", + "choice_c": "trumpet", + "choice_d": "guitar", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18068 + }, + { + "path": "882514.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "electricpiano", + "choice_b": "synthesizer", + "choice_c": "guitar", + "choice_d": "sampler", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18069 + }, + { + "path": "1403968.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "drums", + "choice_b": "bass", + "choice_c": "electricpiano", + "choice_d": "orchestra", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18070 + }, + { + "path": "1047962.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "synthesizer", + "choice_b": "orchestra", + "choice_c": "saxophone", + "choice_d": "strings", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18071 + }, + { + "path": "1374145.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "drummachine", + "choice_b": "electricguitar", + "choice_c": "saxophone", + "choice_d": "rhodes", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18072 + }, + { + "path": "1241457.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "harmonica", + "choice_b": "saxophone", + "choice_c": "flute", + "choice_d": "clarinet", + "answer_gt": "clarinet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18073 + }, + { + "path": "1280441.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "acousticguitar", + "choice_b": "clarinet", + "choice_c": "synthesizer", + "choice_d": "electricguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18074 + }, + { + "path": "207733.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "organ", + "choice_b": "horn", + "choice_c": "pipeorgan", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18075 + }, + { + "path": "1348924.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "synthesizer", + "choice_b": "voice", + "choice_c": "keyboard", + "choice_d": "piano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18076 + }, + { + "path": "25218.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "bell", + "choice_b": "orchestra", + "choice_c": "clarinet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18077 + }, + { + "path": "1126846.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "harp", + "choice_b": "synthesizer", + "choice_c": "violin", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18078 + }, + { + "path": "321963.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "computer", + "choice_b": "keyboard", + "choice_c": "cello", + "choice_d": "synthesizer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18079 + }, + { + "path": "1293686.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "clarinet", + "choice_b": "beat", + "choice_c": "flute", + "choice_d": "percussion", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18080 + }, + { + "path": "387598.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "acousticguitar", + "choice_b": "keyboard", + "choice_c": "accordion", + "choice_d": "cello", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18081 + }, + { + "path": "1280355.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "flute", + "choice_b": "bell", + "choice_c": "violin", + "choice_d": "saxophone", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18082 + }, + { + "path": "1415222.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "piano", + "choice_b": "rhodes", + "choice_c": "violin", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18083 + }, + { + "path": "1026347.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "synthesizer", + "choice_b": "keyboard", + "choice_c": "horn", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18084 + }, + { + "path": "1026345.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "beat", + "choice_c": "trumpet", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18085 + }, + { + "path": "1124711.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "electricguitar", + "choice_b": "percussion", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18086 + }, + { + "path": "1160114.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "classicalguitar", + "choice_b": "computer", + "choice_c": "rhodes", + "choice_d": "harp", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18087 + }, + { + "path": "1302905.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "drummachine", + "choice_b": "piano", + "choice_c": "electricpiano", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18088 + }, + { + "path": "25242.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "trumpet", + "choice_b": "pipeorgan", + "choice_c": "guitar", + "choice_d": "classicalguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18089 + }, + { + "path": "1219428.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "horn", + "choice_b": "bell", + "choice_c": "electricpiano", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18090 + }, + { + "path": "1103549.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "pad", + "choice_b": "beat", + "choice_c": "synthesizer", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18091 + }, + { + "path": "1196911.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "saxophone", + "choice_b": "brass", + "choice_c": "computer", + "choice_d": "sampler", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18092 + }, + { + "path": "1174155.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "pipeorgan", + "choice_b": "clarinet", + "choice_c": "synthesizer", + "choice_d": "piano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18093 + }, + { + "path": "1291569.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "percussion", + "choice_b": "piano", + "choice_c": "saxophone", + "choice_d": "organ", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18094 + }, + { + "path": "903604.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "pipeorgan", + "choice_b": "clarinet", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18095 + }, + { + "path": "457519.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "violin", + "choice_b": "harp", + "choice_c": "sampler", + "choice_d": "rhodes", + "answer_gt": "sampler", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18096 + }, + { + "path": "398183.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "classicalguitar", + "choice_b": "accordion", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18097 + }, + { + "path": "76976.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "violin", + "choice_b": "accordion", + "choice_c": "pad", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18098 + }, + { + "path": "1280420.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "electricguitar", + "choice_b": "classicalguitar", + "choice_c": "harmonica", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18099 + }, + { + "path": "138323.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "piano", + "choice_b": "cello", + "choice_c": "drummachine", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18100 + }, + { + "path": "511052.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "flute", + "choice_b": "electricguitar", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18101 + }, + { + "path": "881055.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "synthesizer", + "choice_b": "voice", + "choice_c": "clarinet", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18102 + }, + { + "path": "979841.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "synthesizer", + "choice_b": "brass", + "choice_c": "pad", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18103 + }, + { + "path": "816202.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "trumpet", + "choice_b": "synthesizer", + "choice_c": "piano", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18104 + }, + { + "path": "850066.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "percussion", + "choice_b": "bass", + "choice_c": "classicalguitar", + "choice_d": "drums", + "answer_gt": "percussion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18105 + }, + { + "path": "1342574.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "violin", + "choice_b": "accordion", + "choice_c": "harp", + "choice_d": "pad", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18106 + }, + { + "path": "577325.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "rhodes", + "choice_b": "harp", + "choice_c": "electricguitar", + "choice_d": "computer", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18107 + }, + { + "path": "1163962.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "beat", + "choice_c": "cello", + "choice_d": "bass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18108 + }, + { + "path": "1264980.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "accordion", + "choice_b": "pad", + "choice_c": "bell", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18109 + }, + { + "path": "80805.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "saxophone", + "choice_b": "electricguitar", + "choice_c": "computer", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18110 + }, + { + "path": "1219425.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "flute", + "choice_c": "voice", + "choice_d": "guitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18111 + }, + { + "path": "903591.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "guitar", + "choice_b": "piano", + "choice_c": "clarinet", + "choice_d": "accordion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18112 + }, + { + "path": "1232309.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "cello", + "choice_c": "piano", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18113 + }, + { + "path": "1203872.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "piano", + "choice_b": "flute", + "choice_c": "classicalguitar", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18114 + }, + { + "path": "584804.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "harmonica", + "choice_b": "drums", + "choice_c": "guitar", + "choice_d": "pad", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18115 + }, + { + "path": "1243739.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "pipeorgan", + "choice_c": "beat", + "choice_d": "cello", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18116 + }, + { + "path": "1076562.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "electricpiano", + "choice_b": "beat", + "choice_c": "brass", + "choice_d": "keyboard", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18117 + }, + { + "path": "1326219.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "flute", + "choice_b": "electricpiano", + "choice_c": "acousticguitar", + "choice_d": "percussion", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18118 + }, + { + "path": "1244443.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "pad", + "choice_b": "bass", + "choice_c": "strings", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18119 + }, + { + "path": "149491.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "voice", + "choice_b": "drummachine", + "choice_c": "harmonica", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18120 + }, + { + "path": "1257403.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "pipeorgan", + "choice_d": "percussion", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18121 + }, + { + "path": "1000082.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "pad", + "choice_b": "acousticguitar", + "choice_c": "orchestra", + "choice_d": "strings", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18122 + }, + { + "path": "949555.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "cello", + "choice_b": "horn", + "choice_c": "classicalguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18123 + }, + { + "path": "1348114.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "accordion", + "choice_b": "drums", + "choice_c": "piano", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18124 + }, + { + "path": "243697.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "harp", + "choice_b": "acousticguitar", + "choice_c": "voice", + "choice_d": "brass", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18125 + }, + { + "path": "1179522.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "orchestra", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18126 + }, + { + "path": "1164343.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "trumpet", + "choice_b": "acousticguitar", + "choice_c": "strings", + "choice_d": "guitar", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18127 + }, + { + "path": "1188947.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "piano", + "choice_b": "saxophone", + "choice_c": "acousticguitar", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18128 + }, + { + "path": "1178110.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "violin", + "choice_b": "saxophone", + "choice_c": "synthesizer", + "choice_d": "harp", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18129 + }, + { + "path": "1347850.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "synthesizer", + "choice_b": "guitar", + "choice_c": "bell", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18130 + }, + { + "path": "502325.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "acousticguitar", + "choice_b": "piano", + "choice_c": "accordion", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18131 + }, + { + "path": "243809.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "voice", + "choice_b": "guitar", + "choice_c": "pipeorgan", + "choice_d": "trumpet", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18132 + }, + { + "path": "1301595.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "guitar", + "choice_b": "piano", + "choice_c": "bass", + "choice_d": "voice", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18133 + }, + { + "path": "94825.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "acousticguitar", + "choice_b": "cello", + "choice_c": "piano", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18134 + }, + { + "path": "1418668.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "guitar", + "choice_b": "pad", + "choice_c": "acousticguitar", + "choice_d": "pipeorgan", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18135 + }, + { + "path": "858537.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "clarinet", + "choice_b": "piano", + "choice_c": "saxophone", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18136 + }, + { + "path": "1225866.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "classicalguitar", + "choice_b": "acousticguitar", + "choice_c": "pad", + "choice_d": "computer", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18137 + }, + { + "path": "1131679.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "trumpet", + "choice_b": "computer", + "choice_c": "voice", + "choice_d": "saxophone", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18138 + }, + { + "path": "30223.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "harmonica", + "choice_c": "guitar", + "choice_d": "trumpet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18139 + }, + { + "path": "243721.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "rhodes", + "choice_b": "voice", + "choice_c": "bell", + "choice_d": "accordion", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18140 + }, + { + "path": "1371515.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "clarinet", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18141 + }, + { + "path": "1204956.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "electricpiano", + "choice_b": "beat", + "choice_c": "guitar", + "choice_d": "acousticguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18142 + }, + { + "path": "376419.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "keyboard", + "choice_b": "rhodes", + "choice_c": "acousticguitar", + "choice_d": "electricpiano", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18143 + }, + { + "path": "82638.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "piano", + "choice_b": "horn", + "choice_c": "synthesizer", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18144 + }, + { + "path": "1117572.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "synthesizer", + "choice_b": "strings", + "choice_c": "bass", + "choice_d": "trumpet", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18145 + }, + { + "path": "866705.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "strings", + "choice_b": "violin", + "choice_c": "orchestra", + "choice_d": "accordion", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18146 + }, + { + "path": "1204667.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "bell", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "harp", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18147 + }, + { + "path": "1203426.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "drummachine", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "pipeorgan", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18148 + }, + { + "path": "1365991.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "electricpiano", + "choice_b": "violin", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18149 + }, + { + "path": "1054172.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "electricpiano", + "choice_b": "strings", + "choice_c": "classicalguitar", + "choice_d": "synthesizer", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18150 + }, + { + "path": "1085249.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "violin", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "accordion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18151 + }, + { + "path": "1313333.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "beat", + "choice_b": "drummachine", + "choice_c": "computer", + "choice_d": "bass", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18152 + }, + { + "path": "103296.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "classicalguitar", + "choice_d": "clarinet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18153 + }, + { + "path": "1294938.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "voice", + "choice_b": "orchestra", + "choice_c": "horn", + "choice_d": "clarinet", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18154 + }, + { + "path": "1312101.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "violin", + "choice_b": "voice", + "choice_c": "guitar", + "choice_d": "sampler", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18155 + }, + { + "path": "57861.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "piano", + "choice_b": "accordion", + "choice_c": "computer", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18156 + }, + { + "path": "1133086.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "flute", + "choice_b": "synthesizer", + "choice_c": "drummachine", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18157 + }, + { + "path": "1241452.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "pad", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18158 + }, + { + "path": "858545.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "piano", + "choice_b": "accordion", + "choice_c": "classicalguitar", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18159 + }, + { + "path": "1083515.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "classicalguitar", + "choice_b": "pipeorgan", + "choice_c": "acousticguitar", + "choice_d": "pad", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18160 + }, + { + "path": "915277.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "piano", + "choice_b": "brass", + "choice_c": "harmonica", + "choice_d": "drummachine", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18161 + }, + { + "path": "398188.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "clarinet", + "choice_b": "pad", + "choice_c": "voice", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18162 + }, + { + "path": "1023485.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "sampler", + "choice_b": "bell", + "choice_c": "violin", + "choice_d": "organ", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18163 + }, + { + "path": "1280434.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "strings", + "choice_b": "synthesizer", + "choice_c": "beat", + "choice_d": "electricguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18164 + }, + { + "path": "1164346.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "strings", + "choice_b": "drummachine", + "choice_c": "guitar", + "choice_d": "beat", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18165 + }, + { + "path": "1128825.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "strings", + "choice_b": "brass", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18166 + }, + { + "path": "1296782.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "brass", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18167 + }, + { + "path": "1416198.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "harp", + "choice_b": "drums", + "choice_c": "classicalguitar", + "choice_d": "cello", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18168 + }, + { + "path": "1076568.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "beat", + "choice_b": "drums", + "choice_c": "percussion", + "choice_d": "piano", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18169 + }, + { + "path": "368162.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "violin", + "choice_b": "percussion", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18170 + }, + { + "path": "1076579.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "drummachine", + "choice_b": "rhodes", + "choice_c": "beat", + "choice_d": "orchestra", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18171 + }, + { + "path": "1418114.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "strings", + "choice_b": "horn", + "choice_c": "beat", + "choice_d": "orchestra", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18172 + }, + { + "path": "1004129.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "violin", + "choice_b": "beat", + "choice_c": "saxophone", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18173 + }, + { + "path": "1076563.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "bell", + "choice_b": "beat", + "choice_c": "classicalguitar", + "choice_d": "cello", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18174 + }, + { + "path": "1301712.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "harp", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "harmonica", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18175 + }, + { + "path": "915292.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "accordion", + "choice_b": "saxophone", + "choice_c": "drummachine", + "choice_d": "horn", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18176 + }, + { + "path": "368179.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "piano", + "choice_b": "orchestra", + "choice_c": "computer", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18177 + }, + { + "path": "1300020.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "keyboard", + "choice_b": "guitar", + "choice_c": "synthesizer", + "choice_d": "brass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18178 + }, + { + "path": "80795.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "pad", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18179 + }, + { + "path": "1227499.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "accordion", + "choice_c": "electricpiano", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18180 + }, + { + "path": "1160226.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "piano", + "choice_b": "bass", + "choice_c": "classicalguitar", + "choice_d": "organ", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18181 + }, + { + "path": "330212.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "piano", + "choice_b": "keyboard", + "choice_c": "voice", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18182 + }, + { + "path": "1012003.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "rhodes", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18183 + }, + { + "path": "1360757.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "flute", + "choice_b": "cello", + "choice_c": "bell", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18184 + }, + { + "path": "923522.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "strings", + "choice_b": "synthesizer", + "choice_c": "organ", + "choice_d": "pad", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18185 + }, + { + "path": "687986.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "rhodes", + "choice_b": "electricpiano", + "choice_c": "organ", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18186 + }, + { + "path": "1311817.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "horn", + "choice_b": "drummachine", + "choice_c": "piano", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18187 + }, + { + "path": "25222.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "accordion", + "choice_b": "piano", + "choice_c": "harmonica", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18188 + }, + { + "path": "1116412.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "classicalguitar", + "choice_b": "bell", + "choice_c": "pipeorgan", + "choice_d": "electricguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18189 + }, + { + "path": "1314853.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "guitar", + "choice_b": "cello", + "choice_c": "rhodes", + "choice_d": "classicalguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18190 + }, + { + "path": "1300001.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "clarinet", + "choice_b": "accordion", + "choice_c": "synthesizer", + "choice_d": "bass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18191 + }, + { + "path": "664813.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "beat", + "choice_b": "computer", + "choice_c": "pipeorgan", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18192 + }, + { + "path": "1403969.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "clarinet", + "choice_b": "piano", + "choice_c": "brass", + "choice_d": "trumpet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18193 + }, + { + "path": "1076575.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "flute", + "choice_c": "synthesizer", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18194 + }, + { + "path": "1119923.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "clarinet", + "choice_b": "rhodes", + "choice_c": "classicalguitar", + "choice_d": "computer", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18195 + }, + { + "path": "1100153.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "orchestra", + "choice_b": "accordion", + "choice_c": "computer", + "choice_d": "pipeorgan", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18196 + }, + { + "path": "23869.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "drums", + "choice_b": "guitar", + "choice_c": "strings", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18197 + }, + { + "path": "1031108.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "beat", + "choice_c": "saxophone", + "choice_d": "electricguitar", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18198 + }, + { + "path": "1243729.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "beat", + "choice_b": "acousticguitar", + "choice_c": "bell", + "choice_d": "drummachine", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18199 + }, + { + "path": "1219426.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "bass", + "choice_b": "synthesizer", + "choice_c": "organ", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18200 + }, + { + "path": "176084.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "harp", + "choice_b": "bass", + "choice_c": "clarinet", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18201 + }, + { + "path": "1310775.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "guitar", + "choice_b": "bell", + "choice_c": "strings", + "choice_d": "drummachine", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18202 + }, + { + "path": "1179510.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "rhodes", + "choice_b": "classicalguitar", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18203 + }, + { + "path": "799915.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "organ", + "choice_b": "voice", + "choice_c": "classicalguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18204 + }, + { + "path": "873136.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "acousticguitar", + "choice_b": "accordion", + "choice_c": "piano", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18205 + }, + { + "path": "1371512.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "bass", + "choice_b": "harmonica", + "choice_c": "guitar", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18206 + }, + { + "path": "1294933.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "voice", + "choice_b": "trumpet", + "choice_c": "beat", + "choice_d": "organ", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18207 + }, + { + "path": "1291558.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "brass", + "choice_b": "strings", + "choice_c": "piano", + "choice_d": "flute", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18208 + }, + { + "path": "1163775.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "harp", + "choice_b": "voice", + "choice_c": "synthesizer", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18209 + }, + { + "path": "686358.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "classicalguitar", + "choice_b": "piano", + "choice_c": "organ", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18210 + }, + { + "path": "80796.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "keyboard", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18211 + }, + { + "path": "1268986.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "bass", + "choice_b": "brass", + "choice_c": "cello", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18212 + }, + { + "path": "1204670.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "saxophone", + "choice_b": "piano", + "choice_c": "guitar", + "choice_d": "voice", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18213 + }, + { + "path": "969349.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "violin", + "choice_b": "pipeorgan", + "choice_c": "piano", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18214 + }, + { + "path": "1203669.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "harmonica", + "choice_b": "guitar", + "choice_c": "classicalguitar", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18215 + }, + { + "path": "1246338.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "acousticguitar", + "choice_b": "electricguitar", + "choice_c": "pad", + "choice_d": "accordion", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18216 + }, + { + "path": "1253537.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "clarinet", + "choice_b": "drummachine", + "choice_c": "bell", + "choice_d": "bass", + "answer_gt": "bell", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18217 + }, + { + "path": "1077724.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "drummachine", + "choice_b": "harp", + "choice_c": "keyboard", + "choice_d": "electricguitar", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18218 + }, + { + "path": "840599.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "organ", + "choice_b": "classicalguitar", + "choice_c": "guitar", + "choice_d": "acousticguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18219 + }, + { + "path": "1356515.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "percussion", + "choice_b": "trumpet", + "choice_c": "flute", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18220 + }, + { + "path": "1225956.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "organ", + "choice_b": "acousticguitar", + "choice_c": "keyboard", + "choice_d": "bass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18221 + }, + { + "path": "80786.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "piano", + "choice_b": "computer", + "choice_c": "strings", + "choice_d": "pad", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18222 + }, + { + "path": "442223.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bass", + "choice_b": "piano", + "choice_c": "keyboard", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18223 + }, + { + "path": "1121556.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "electricguitar", + "choice_b": "accordion", + "choice_c": "bass", + "choice_d": "flute", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18224 + }, + { + "path": "259320.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "beat", + "choice_b": "piano", + "choice_c": "acousticguitar", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18225 + }, + { + "path": "1031306.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "electricguitar", + "choice_c": "trumpet", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18226 + }, + { + "path": "1117701.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "piano", + "choice_c": "flute", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18227 + }, + { + "path": "589410.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "acousticguitar", + "choice_b": "trumpet", + "choice_c": "orchestra", + "choice_d": "clarinet", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18228 + }, + { + "path": "1133081.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "orchestra", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18229 + }, + { + "path": "1395662.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "strings", + "choice_b": "classicalguitar", + "choice_c": "electricpiano", + "choice_d": "pad", + "answer_gt": "pad", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18230 + }, + { + "path": "25237.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "beat", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18231 + }, + { + "path": "384144.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "trumpet", + "choice_b": "guitar", + "choice_c": "violin", + "choice_d": "horn", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18232 + }, + { + "path": "1076574.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "beat", + "choice_b": "classicalguitar", + "choice_c": "pipeorgan", + "choice_d": "bass", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18233 + }, + { + "path": "1280923.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "computer", + "choice_b": "pipeorgan", + "choice_c": "voice", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18234 + }, + { + "path": "1356470.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "guitar", + "choice_b": "electricpiano", + "choice_c": "acousticguitar", + "choice_d": "accordion", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18235 + }, + { + "path": "1226206.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "harmonica", + "choice_b": "organ", + "choice_c": "acousticguitar", + "choice_d": "synthesizer", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18236 + }, + { + "path": "41977.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "synthesizer", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18237 + }, + { + "path": "1128817.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "drums", + "choice_b": "organ", + "choice_c": "keyboard", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18238 + }, + { + "path": "1128823.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "computer", + "choice_b": "piano", + "choice_c": "classicalguitar", + "choice_d": "horn", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18239 + }, + { + "path": "824624.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "beat", + "choice_b": "synthesizer", + "choice_c": "harp", + "choice_d": "horn", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18240 + }, + { + "path": "1133083.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "piano", + "choice_b": "violin", + "choice_c": "keyboard", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18241 + }, + { + "path": "1163803.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "orchestra", + "choice_b": "beat", + "choice_c": "organ", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18242 + }, + { + "path": "1244715.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "violin", + "choice_b": "brass", + "choice_c": "synthesizer", + "choice_d": "organ", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18243 + }, + { + "path": "368174.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "strings", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18244 + }, + { + "path": "243794.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "accordion", + "choice_b": "electricpiano", + "choice_c": "voice", + "choice_d": "violin", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18245 + }, + { + "path": "467716.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "accordion", + "choice_b": "horn", + "choice_c": "piano", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18246 + }, + { + "path": "1102806.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "acousticguitar", + "choice_b": "rhodes", + "choice_c": "harp", + "choice_d": "drums", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18247 + }, + { + "path": "1077718.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "pipeorgan", + "choice_b": "synthesizer", + "choice_c": "trumpet", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18248 + }, + { + "path": "511090.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "flute", + "choice_b": "harp", + "choice_c": "acousticguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18249 + }, + { + "path": "757498.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "drums", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18250 + }, + { + "path": "1389038.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "guitar", + "choice_b": "drums", + "choice_c": "pad", + "choice_d": "voice", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18251 + }, + { + "path": "1310024.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "piano", + "choice_b": "synthesizer", + "choice_c": "electricpiano", + "choice_d": "classicalguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18252 + }, + { + "path": "1073701.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "brass", + "choice_b": "electricpiano", + "choice_c": "saxophone", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18253 + }, + { + "path": "1036085.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "orchestra", + "choice_b": "flute", + "choice_c": "synthesizer", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18254 + }, + { + "path": "1395311.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "computer", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "horn", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18255 + }, + { + "path": "1203871.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "horn", + "choice_b": "computer", + "choice_c": "piano", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18256 + }, + { + "path": "677105.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "harmonica", + "choice_b": "synthesizer", + "choice_c": "acousticguitar", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18257 + }, + { + "path": "1164746.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "rhodes", + "choice_b": "harp", + "choice_c": "orchestra", + "choice_d": "synthesizer", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18258 + }, + { + "path": "398187.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "percussion", + "choice_b": "drummachine", + "choice_c": "sampler", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18259 + }, + { + "path": "1032644.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "classicalguitar", + "choice_b": "brass", + "choice_c": "acousticguitar", + "choice_d": "sampler", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18260 + }, + { + "path": "373053.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "piano", + "choice_b": "strings", + "choice_c": "brass", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18261 + }, + { + "path": "979846.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "keyboard", + "choice_b": "synthesizer", + "choice_c": "strings", + "choice_d": "organ", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18262 + }, + { + "path": "1029507.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "synthesizer", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18263 + }, + { + "path": "1086691.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "violin", + "choice_c": "voice", + "choice_d": "pipeorgan", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18264 + }, + { + "path": "824636.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "electricpiano", + "choice_b": "synthesizer", + "choice_c": "clarinet", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18265 + }, + { + "path": "1225957.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "electricpiano", + "choice_b": "harp", + "choice_c": "acousticguitar", + "choice_d": "brass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18266 + }, + { + "path": "1164345.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "rhodes", + "choice_b": "bass", + "choice_c": "voice", + "choice_d": "strings", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18267 + }, + { + "path": "1053502.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "clarinet", + "choice_b": "electricpiano", + "choice_c": "flute", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18268 + }, + { + "path": "1160047.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "acousticguitar", + "choice_b": "orchestra", + "choice_c": "classicalguitar", + "choice_d": "bass", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18269 + }, + { + "path": "131618.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "orchestra", + "choice_b": "bass", + "choice_c": "guitar", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18270 + }, + { + "path": "862191.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "bell", + "choice_b": "horn", + "choice_c": "piano", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18271 + }, + { + "path": "38114.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "synthesizer", + "choice_b": "keyboard", + "choice_c": "drums", + "choice_d": "percussion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18272 + }, + { + "path": "1062497.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "electricpiano", + "choice_b": "beat", + "choice_c": "orchestra", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18273 + }, + { + "path": "742095.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "classicalguitar", + "choice_b": "saxophone", + "choice_c": "pipeorgan", + "choice_d": "cello", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18274 + }, + { + "path": "948722.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "horn", + "choice_b": "accordion", + "choice_c": "violin", + "choice_d": "classicalguitar", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18275 + }, + { + "path": "1183239.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "trumpet", + "choice_b": "piano", + "choice_c": "flute", + "choice_d": "pad", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18276 + }, + { + "path": "979837.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "rhodes", + "choice_b": "beat", + "choice_c": "synthesizer", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18277 + }, + { + "path": "1278181.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "percussion", + "choice_b": "horn", + "choice_c": "clarinet", + "choice_d": "orchestra", + "answer_gt": "clarinet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18278 + }, + { + "path": "1333332.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "horn", + "choice_b": "flute", + "choice_c": "guitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18279 + }, + { + "path": "847264.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "violin", + "choice_b": "flute", + "choice_c": "piano", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18280 + }, + { + "path": "944146.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "accordion", + "choice_b": "synthesizer", + "choice_c": "piano", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18281 + }, + { + "path": "903596.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "clarinet", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18282 + }, + { + "path": "1286511.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "pad", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18283 + }, + { + "path": "287980.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "acousticguitar", + "choice_b": "clarinet", + "choice_c": "synthesizer", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18284 + }, + { + "path": "25243.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "bass", + "choice_b": "beat", + "choice_c": "rhodes", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18285 + }, + { + "path": "941857.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "percussion", + "choice_b": "electricguitar", + "choice_c": "computer", + "choice_d": "electricpiano", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18286 + }, + { + "path": "1083512.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "violin", + "choice_b": "harmonica", + "choice_c": "strings", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18287 + }, + { + "path": "1395671.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "pipeorgan", + "choice_b": "piano", + "choice_c": "clarinet", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18288 + }, + { + "path": "951887.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "clarinet", + "choice_b": "synthesizer", + "choice_c": "acousticguitar", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18289 + }, + { + "path": "1319227.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "guitar", + "choice_d": "saxophone", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18290 + }, + { + "path": "988996.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "guitar", + "choice_b": "horn", + "choice_c": "electricpiano", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18291 + }, + { + "path": "702503.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "electricguitar", + "choice_b": "orchestra", + "choice_c": "keyboard", + "choice_d": "percussion", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18292 + }, + { + "path": "589408.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "synthesizer", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18293 + }, + { + "path": "1248622.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "beat", + "choice_b": "accordion", + "choice_c": "orchestra", + "choice_d": "cello", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18294 + }, + { + "path": "1053492.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "computer", + "choice_b": "accordion", + "choice_c": "classicalguitar", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18295 + }, + { + "path": "979848.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "percussion", + "choice_b": "electricpiano", + "choice_c": "synthesizer", + "choice_d": "bell", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18296 + }, + { + "path": "575468.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "classicalguitar", + "choice_b": "acousticguitar", + "choice_c": "horn", + "choice_d": "bass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18297 + }, + { + "path": "257176.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "piano", + "choice_b": "computer", + "choice_c": "percussion", + "choice_d": "trumpet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18298 + }, + { + "path": "36984.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "violin", + "choice_b": "flute", + "choice_c": "sampler", + "choice_d": "harp", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18299 + }, + { + "path": "1310774.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "guitar", + "choice_b": "orchestra", + "choice_c": "rhodes", + "choice_d": "pipeorgan", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18300 + }, + { + "path": "1156898.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "computer", + "choice_b": "sampler", + "choice_c": "pipeorgan", + "choice_d": "bell", + "answer_gt": "pipeorgan", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18301 + }, + { + "path": "1047961.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "synthesizer", + "choice_b": "electricpiano", + "choice_c": "pad", + "choice_d": "percussion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18302 + }, + { + "path": "1037681.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "keyboard", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18303 + }, + { + "path": "1159639.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "classicalguitar", + "choice_b": "drums", + "choice_c": "acousticguitar", + "choice_d": "drummachine", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18304 + }, + { + "path": "1083526.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "accordion", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18305 + }, + { + "path": "824629.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "flute", + "choice_b": "rhodes", + "choice_c": "saxophone", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18306 + }, + { + "path": "1246341.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "electricguitar", + "choice_b": "drums", + "choice_c": "acousticguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18307 + }, + { + "path": "1319239.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "harp", + "choice_b": "synthesizer", + "choice_c": "electricguitar", + "choice_d": "beat", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18308 + }, + { + "path": "1380115.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "cello", + "choice_b": "brass", + "choice_c": "violin", + "choice_d": "bass", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18309 + }, + { + "path": "1117575.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "classicalguitar", + "choice_b": "strings", + "choice_c": "bell", + "choice_d": "drummachine", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18310 + }, + { + "path": "782163.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "computer", + "choice_b": "voice", + "choice_c": "keyboard", + "choice_d": "clarinet", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18311 + }, + { + "path": "184214.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "voice", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18312 + }, + { + "path": "944150.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "keyboard", + "choice_b": "saxophone", + "choice_c": "piano", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18313 + }, + { + "path": "1219360.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "sampler", + "choice_b": "synthesizer", + "choice_c": "electricguitar", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18314 + }, + { + "path": "38275.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "orchestra", + "choice_b": "strings", + "choice_c": "bell", + "choice_d": "harmonica", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18315 + }, + { + "path": "736792.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "harmonica", + "choice_b": "orchestra", + "choice_c": "electricguitar", + "choice_d": "trumpet", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18316 + }, + { + "path": "1274593.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "percussion", + "choice_b": "trumpet", + "choice_c": "harp", + "choice_d": "rhodes", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18317 + }, + { + "path": "277778.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "harmonica", + "choice_b": "guitar", + "choice_c": "acousticguitar", + "choice_d": "drums", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18318 + }, + { + "path": "1280424.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "synthesizer", + "choice_b": "beat", + "choice_c": "drummachine", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18319 + }, + { + "path": "6248.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "harp", + "choice_d": "flute", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18320 + }, + { + "path": "558568.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "accordion", + "choice_c": "orchestra", + "choice_d": "computer", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18321 + }, + { + "path": "1220670.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "harmonica", + "choice_b": "strings", + "choice_c": "orchestra", + "choice_d": "saxophone", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18322 + }, + { + "path": "1418294.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "electricguitar", + "choice_b": "beat", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18323 + }, + { + "path": "1250855.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "harmonica", + "choice_b": "cello", + "choice_c": "keyboard", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18324 + }, + { + "path": "1301828.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "pad", + "choice_c": "synthesizer", + "choice_d": "electricguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18325 + }, + { + "path": "1221322.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "flute", + "choice_b": "bass", + "choice_c": "drummachine", + "choice_d": "classicalguitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18326 + }, + { + "path": "1374331.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "pad", + "choice_b": "flute", + "choice_c": "electricpiano", + "choice_d": "classicalguitar", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18327 + }, + { + "path": "979853.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "organ", + "choice_b": "electricguitar", + "choice_c": "synthesizer", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18328 + }, + { + "path": "1119819.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "piano", + "choice_b": "cello", + "choice_c": "pad", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18329 + }, + { + "path": "919995.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "saxophone", + "choice_b": "drummachine", + "choice_c": "violin", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18330 + }, + { + "path": "1062498.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "bass", + "choice_b": "synthesizer", + "choice_c": "trumpet", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18331 + }, + { + "path": "1168549.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "violin", + "choice_b": "electricpiano", + "choice_c": "saxophone", + "choice_d": "clarinet", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18332 + }, + { + "path": "1203865.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "percussion", + "choice_c": "piano", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18333 + }, + { + "path": "80783.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "percussion", + "choice_b": "harp", + "choice_c": "piano", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18334 + }, + { + "path": "1053491.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "organ", + "choice_b": "harmonica", + "choice_c": "bass", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18335 + }, + { + "path": "1131678.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "voice", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "cello", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18336 + }, + { + "path": "1053494.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "acousticguitar", + "choice_b": "brass", + "choice_c": "computer", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18337 + }, + { + "path": "477312.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "orchestra", + "choice_b": "cello", + "choice_c": "synthesizer", + "choice_d": "strings", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18338 + }, + { + "path": "1395102.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "flute", + "choice_b": "clarinet", + "choice_c": "classicalguitar", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18339 + }, + { + "path": "1284784.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "electricguitar", + "choice_b": "rhodes", + "choice_c": "beat", + "choice_d": "flute", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18340 + }, + { + "path": "108751.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "beat", + "choice_d": "horn", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18341 + }, + { + "path": "1192549.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "pipeorgan", + "choice_b": "electricpiano", + "choice_c": "classicalguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18342 + }, + { + "path": "944149.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "percussion", + "choice_b": "piano", + "choice_c": "beat", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18343 + }, + { + "path": "1356519.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "harmonica", + "choice_b": "flute", + "choice_c": "synthesizer", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18344 + }, + { + "path": "1133074.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "harmonica", + "choice_b": "voice", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18345 + }, + { + "path": "1311816.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "computer", + "choice_b": "piano", + "choice_c": "harmonica", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18346 + }, + { + "path": "1294442.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "synthesizer", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "harp", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18347 + }, + { + "path": "1306027.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "cello", + "choice_b": "drums", + "choice_c": "bass", + "choice_d": "orchestra", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18348 + }, + { + "path": "1362287.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "harp", + "choice_c": "voice", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18349 + }, + { + "path": "1227222.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "computer", + "choice_b": "percussion", + "choice_c": "rhodes", + "choice_d": "keyboard", + "answer_gt": "rhodes", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18350 + }, + { + "path": "1085878.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "bell", + "choice_b": "acousticguitar", + "choice_c": "strings", + "choice_d": "electricguitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18351 + }, + { + "path": "686359.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "piano", + "choice_b": "classicalguitar", + "choice_c": "voice", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18352 + }, + { + "path": "1279177.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "computer", + "choice_b": "saxophone", + "choice_c": "clarinet", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18353 + }, + { + "path": "1395646.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "electricguitar", + "choice_b": "classicalguitar", + "choice_c": "orchestra", + "choice_d": "strings", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18354 + }, + { + "path": "911357.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "synthesizer", + "choice_b": "beat", + "choice_c": "cello", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18355 + }, + { + "path": "1093638.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "brass", + "choice_b": "voice", + "choice_c": "bell", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18356 + }, + { + "path": "323679.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "computer", + "choice_b": "drummachine", + "choice_c": "clarinet", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18357 + }, + { + "path": "824628.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "cello", + "choice_c": "sampler", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18358 + }, + { + "path": "1237953.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "percussion", + "choice_c": "sampler", + "choice_d": "clarinet", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18359 + }, + { + "path": "1053493.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "flute", + "choice_b": "electricpiano", + "choice_c": "synthesizer", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18360 + }, + { + "path": "375489.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "harmonica", + "choice_b": "clarinet", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18361 + }, + { + "path": "1232307.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "orchestra", + "choice_b": "piano", + "choice_c": "harp", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18362 + }, + { + "path": "1096872.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "classicalguitar", + "choice_b": "harp", + "choice_c": "keyboard", + "choice_d": "cello", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18363 + }, + { + "path": "477305.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "percussion", + "choice_b": "orchestra", + "choice_c": "brass", + "choice_d": "computer", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18364 + }, + { + "path": "24841.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "clarinet", + "choice_b": "violin", + "choice_c": "beat", + "choice_d": "flute", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18365 + }, + { + "path": "1039627.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "acousticguitar", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18366 + }, + { + "path": "1245459.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "piano", + "choice_b": "rhodes", + "choice_c": "percussion", + "choice_d": "harmonica", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18367 + }, + { + "path": "1347851.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "brass", + "choice_b": "pipeorgan", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18368 + }, + { + "path": "609174.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "drums", + "choice_b": "percussion", + "choice_c": "sampler", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18369 + }, + { + "path": "1248619.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "beat", + "choice_b": "computer", + "choice_c": "saxophone", + "choice_d": "drummachine", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18370 + }, + { + "path": "1196182.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "organ", + "choice_b": "pad", + "choice_c": "strings", + "choice_d": "synthesizer", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18371 + }, + { + "path": "80774.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "computer", + "choice_b": "brass", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18372 + }, + { + "path": "14014.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "voice", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18373 + }, + { + "path": "1168543.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "bass", + "choice_d": "acousticguitar", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18374 + }, + { + "path": "25224.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "saxophone", + "choice_b": "percussion", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18375 + }, + { + "path": "243856.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "pipeorgan", + "choice_b": "percussion", + "choice_c": "voice", + "choice_d": "electricguitar", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18376 + }, + { + "path": "284170.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "violin", + "choice_b": "rhodes", + "choice_c": "organ", + "choice_d": "voice", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18377 + }, + { + "path": "1077038.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "harp", + "choice_b": "keyboard", + "choice_c": "orchestra", + "choice_d": "organ", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18378 + }, + { + "path": "384148.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "synthesizer", + "choice_d": "harp", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18379 + }, + { + "path": "1374147.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "flute", + "choice_b": "bell", + "choice_c": "synthesizer", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18380 + }, + { + "path": "655027.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "voice", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18381 + }, + { + "path": "1064763.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "accordion", + "choice_b": "cello", + "choice_c": "saxophone", + "choice_d": "pad", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18382 + }, + { + "path": "1244710.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "synthesizer", + "choice_b": "clarinet", + "choice_c": "saxophone", + "choice_d": "guitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18383 + }, + { + "path": "1257406.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "trumpet", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18384 + }, + { + "path": "1163806.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "piano", + "choice_b": "brass", + "choice_c": "electricpiano", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18385 + }, + { + "path": "828182.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "computer", + "choice_b": "voice", + "choice_c": "flute", + "choice_d": "pipeorgan", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18386 + }, + { + "path": "1203868.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "piano", + "choice_b": "electricpiano", + "choice_c": "bass", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18387 + }, + { + "path": "1416193.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "pad", + "choice_b": "synthesizer", + "choice_c": "guitar", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18388 + }, + { + "path": "82644.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "voice", + "choice_b": "pad", + "choice_c": "flute", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18389 + }, + { + "path": "915289.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "beat", + "choice_b": "horn", + "choice_c": "classicalguitar", + "choice_d": "drummachine", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18390 + }, + { + "path": "858547.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "piano", + "choice_b": "electricguitar", + "choice_c": "brass", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18391 + }, + { + "path": "1293685.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "keyboard", + "choice_b": "rhodes", + "choice_c": "flute", + "choice_d": "acousticguitar", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18392 + }, + { + "path": "1395668.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "piano", + "choice_b": "flute", + "choice_c": "beat", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18393 + }, + { + "path": "1418658.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "guitar", + "choice_b": "electricpiano", + "choice_c": "pad", + "choice_d": "sampler", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18394 + }, + { + "path": "1288117.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "violin", + "choice_b": "organ", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18395 + }, + { + "path": "2971.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "horn", + "choice_b": "sampler", + "choice_c": "guitar", + "choice_d": "pad", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18396 + }, + { + "path": "945024.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "brass", + "choice_b": "flute", + "choice_c": "orchestra", + "choice_d": "harp", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18397 + }, + { + "path": "1159861.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "guitar", + "choice_b": "rhodes", + "choice_c": "piano", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18398 + }, + { + "path": "979860.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "sampler", + "choice_b": "accordion", + "choice_c": "harp", + "choice_d": "bass", + "answer_gt": "sampler", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18399 + }, + { + "path": "370357.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "violin", + "choice_b": "computer", + "choice_c": "strings", + "choice_d": "pipeorgan", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18400 + }, + { + "path": "1279774.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "organ", + "choice_b": "orchestra", + "choice_c": "guitar", + "choice_d": "classicalguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18401 + }, + { + "path": "1210889.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "pipeorgan", + "choice_b": "bass", + "choice_c": "classicalguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18402 + }, + { + "path": "1203672.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "guitar", + "choice_b": "voice", + "choice_c": "harmonica", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18403 + }, + { + "path": "1053503.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "voice", + "choice_c": "flute", + "choice_d": "pad", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18404 + }, + { + "path": "1104259.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "orchestra", + "choice_b": "piano", + "choice_c": "harmonica", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18405 + }, + { + "path": "24842.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "strings", + "choice_b": "piano", + "choice_c": "computer", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18406 + }, + { + "path": "1310777.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "trumpet", + "choice_b": "saxophone", + "choice_c": "guitar", + "choice_d": "classicalguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18407 + }, + { + "path": "998389.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "electricguitar", + "choice_b": "cello", + "choice_c": "harmonica", + "choice_d": "percussion", + "answer_gt": "cello", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18408 + }, + { + "path": "1168544.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "electricpiano", + "choice_b": "piano", + "choice_c": "electricguitar", + "choice_d": "harmonica", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18409 + }, + { + "path": "457515.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "sampler", + "choice_b": "strings", + "choice_c": "classicalguitar", + "choice_d": "drums", + "answer_gt": "sampler", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18410 + }, + { + "path": "1370658.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "percussion", + "choice_b": "violin", + "choice_c": "guitar", + "choice_d": "harmonica", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18411 + }, + { + "path": "1280432.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "synthesizer", + "choice_b": "guitar", + "choice_c": "acousticguitar", + "choice_d": "electricguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18412 + }, + { + "path": "979849.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "saxophone", + "choice_b": "synthesizer", + "choice_c": "keyboard", + "choice_d": "brass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18413 + }, + { + "path": "106895.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "classicalguitar", + "choice_c": "flute", + "choice_d": "drummachine", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18414 + }, + { + "path": "368187.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "violin", + "choice_b": "harp", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18415 + }, + { + "path": "14015.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "classicalguitar", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18416 + }, + { + "path": "1164347.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "harp", + "choice_b": "strings", + "choice_c": "synthesizer", + "choice_d": "clarinet", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18417 + }, + { + "path": "23865.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "horn", + "choice_b": "drums", + "choice_c": "harmonica", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18418 + }, + { + "path": "1083528.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "electricguitar", + "choice_b": "piano", + "choice_c": "strings", + "choice_d": "classicalguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18419 + }, + { + "path": "921509.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "saxophone", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "strings", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18420 + }, + { + "path": "80779.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "piano", + "choice_b": "beat", + "choice_c": "classicalguitar", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18421 + }, + { + "path": "204687.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "piano", + "choice_b": "horn", + "choice_c": "acousticguitar", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18422 + }, + { + "path": "80778.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "keyboard", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18423 + }, + { + "path": "131615.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "violin", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18424 + }, + { + "path": "1062499.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "beat", + "choice_b": "acousticguitar", + "choice_c": "violin", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18425 + }, + { + "path": "1280492.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "drums", + "choice_b": "saxophone", + "choice_c": "synthesizer", + "choice_d": "keyboard", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18426 + }, + { + "path": "1280922.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "pipeorgan", + "choice_b": "accordion", + "choice_c": "computer", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18427 + }, + { + "path": "283214.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "piano", + "choice_b": "harmonica", + "choice_c": "horn", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18428 + }, + { + "path": "1216968.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "guitar", + "choice_b": "saxophone", + "choice_c": "strings", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18429 + }, + { + "path": "1225307.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "drums", + "choice_b": "flute", + "choice_c": "harp", + "choice_d": "acousticguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18430 + }, + { + "path": "736791.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "beat", + "choice_b": "orchestra", + "choice_c": "harmonica", + "choice_d": "guitar", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18431 + }, + { + "path": "1349028.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "horn", + "choice_b": "synthesizer", + "choice_c": "brass", + "choice_d": "computer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18432 + }, + { + "path": "1045652.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "brass", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18433 + }, + { + "path": "1279452.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "cello", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18434 + }, + { + "path": "1353701.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "horn", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18435 + }, + { + "path": "1399573.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "percussion", + "choice_b": "strings", + "choice_c": "harp", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18436 + }, + { + "path": "740958.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "bell", + "choice_b": "piano", + "choice_c": "rhodes", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18437 + }, + { + "path": "1191323.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "orchestra", + "choice_b": "synthesizer", + "choice_c": "pad", + "choice_d": "guitar", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18438 + }, + { + "path": "1121560.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "pad", + "choice_b": "sampler", + "choice_c": "classicalguitar", + "choice_d": "rhodes", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18439 + }, + { + "path": "1116413.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "pipeorgan", + "choice_b": "rhodes", + "choice_c": "beat", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18440 + }, + { + "path": "277785.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "cello", + "choice_c": "brass", + "choice_d": "voice", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18441 + }, + { + "path": "1098437.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "piano", + "choice_b": "horn", + "choice_c": "sampler", + "choice_d": "violin", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18442 + }, + { + "path": "965750.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "orchestra", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "synthesizer", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18443 + }, + { + "path": "1302264.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "harp", + "choice_b": "piano", + "choice_c": "guitar", + "choice_d": "trumpet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18444 + }, + { + "path": "1086642.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "horn", + "choice_b": "violin", + "choice_c": "organ", + "choice_d": "synthesizer", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18445 + }, + { + "path": "1083534.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "pad", + "choice_b": "sampler", + "choice_c": "piano", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18446 + }, + { + "path": "1133075.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "clarinet", + "choice_b": "brass", + "choice_c": "piano", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18447 + }, + { + "path": "277816.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "pad", + "choice_b": "computer", + "choice_c": "acousticguitar", + "choice_d": "synthesizer", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18448 + }, + { + "path": "1296781.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "piano", + "choice_b": "keyboard", + "choice_c": "orchestra", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18449 + }, + { + "path": "1117702.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "acousticguitar", + "choice_b": "orchestra", + "choice_c": "saxophone", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18450 + }, + { + "path": "1347951.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "electricguitar", + "choice_c": "accordion", + "choice_d": "pad", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18451 + }, + { + "path": "824630.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "computer", + "choice_b": "beat", + "choice_c": "synthesizer", + "choice_d": "orchestra", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18452 + }, + { + "path": "1210603.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "piano", + "choice_b": "voice", + "choice_c": "harp", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18453 + }, + { + "path": "1266726.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "piano", + "choice_b": "brass", + "choice_c": "synthesizer", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18454 + }, + { + "path": "981255.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "guitar", + "choice_b": "computer", + "choice_c": "harmonica", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18455 + }, + { + "path": "259323.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "piano", + "choice_b": "brass", + "choice_c": "classicalguitar", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18456 + }, + { + "path": "903588.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "classicalguitar", + "choice_b": "piano", + "choice_c": "synthesizer", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18457 + }, + { + "path": "80758.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "bell", + "choice_b": "electricpiano", + "choice_c": "piano", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18458 + }, + { + "path": "1348853.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "voice", + "choice_b": "keyboard", + "choice_c": "computer", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18459 + }, + { + "path": "1349088.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "strings", + "choice_b": "harp", + "choice_c": "horn", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18460 + }, + { + "path": "74752.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "violin", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18461 + }, + { + "path": "1374326.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "horn", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "harmonica", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18462 + }, + { + "path": "1203892.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "acousticguitar", + "choice_b": "clarinet", + "choice_c": "piano", + "choice_d": "pipeorgan", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18463 + }, + { + "path": "243839.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "electricguitar", + "choice_b": "trumpet", + "choice_c": "voice", + "choice_d": "beat", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18464 + }, + { + "path": "1313334.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "pipeorgan", + "choice_b": "beat", + "choice_c": "classicalguitar", + "choice_d": "computer", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18465 + }, + { + "path": "253757.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "saxophone", + "choice_b": "flute", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18466 + }, + { + "path": "1159912.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "synthesizer", + "choice_b": "organ", + "choice_c": "cello", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18467 + }, + { + "path": "979861.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "organ", + "choice_b": "pipeorgan", + "choice_c": "bell", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18468 + }, + { + "path": "1403980.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "horn", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18469 + }, + { + "path": "1163971.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "organ", + "choice_b": "keyboard", + "choice_c": "pad", + "choice_d": "acousticguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18470 + }, + { + "path": "1414803.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "accordion", + "choice_b": "acousticguitar", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18471 + }, + { + "path": "14027.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "keyboard", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18472 + }, + { + "path": "1161202.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "saxophone", + "choice_b": "harmonica", + "choice_c": "drums", + "choice_d": "accordion", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18473 + }, + { + "path": "1248617.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "brass", + "choice_b": "horn", + "choice_c": "harmonica", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18474 + }, + { + "path": "875380.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "keyboard", + "choice_b": "horn", + "choice_c": "synthesizer", + "choice_d": "flute", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18475 + }, + { + "path": "1293688.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "guitar", + "choice_b": "percussion", + "choice_c": "drummachine", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18476 + }, + { + "path": "1284780.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "trumpet", + "choice_b": "keyboard", + "choice_c": "guitar", + "choice_d": "synthesizer", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18477 + }, + { + "path": "361987.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "horn", + "choice_b": "acousticguitar", + "choice_c": "sampler", + "choice_d": "strings", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18478 + }, + { + "path": "1416190.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "guitar", + "choice_b": "synthesizer", + "choice_c": "cello", + "choice_d": "voice", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18479 + }, + { + "path": "1251475.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "strings", + "choice_b": "beat", + "choice_c": "brass", + "choice_d": "harmonica", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18480 + }, + { + "path": "1395310.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "saxophone", + "choice_b": "piano", + "choice_c": "beat", + "choice_d": "electricguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18481 + }, + { + "path": "423237.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "strings", + "choice_b": "classicalguitar", + "choice_c": "drums", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18482 + }, + { + "path": "915275.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "computer", + "choice_b": "harp", + "choice_c": "drummachine", + "choice_d": "cello", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18483 + }, + { + "path": "1315063.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "electricpiano", + "choice_b": "saxophone", + "choice_c": "electricguitar", + "choice_d": "harp", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18484 + }, + { + "path": "184207.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "strings", + "choice_b": "rhodes", + "choice_c": "piano", + "choice_d": "accordion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18485 + }, + { + "path": "1348927.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "horn", + "choice_b": "cello", + "choice_c": "synthesizer", + "choice_d": "harp", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18486 + }, + { + "path": "1391816.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "classicalguitar", + "choice_b": "drums", + "choice_c": "electricguitar", + "choice_d": "horn", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18487 + }, + { + "path": "91820.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "orchestra", + "choice_b": "piano", + "choice_c": "voice", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18488 + }, + { + "path": "862197.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "pipeorgan", + "choice_b": "bass", + "choice_c": "piano", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18489 + }, + { + "path": "919998.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "clarinet", + "choice_b": "violin", + "choice_c": "pad", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18490 + }, + { + "path": "1280924.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "saxophone", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18491 + }, + { + "path": "1348934.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "percussion", + "choice_b": "synthesizer", + "choice_c": "trumpet", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18492 + }, + { + "path": "1374144.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "bass", + "choice_b": "computer", + "choice_c": "synthesizer", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18493 + }, + { + "path": "57862.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "rhodes", + "choice_b": "guitar", + "choice_c": "acousticguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18494 + }, + { + "path": "1083535.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "clarinet", + "choice_b": "piano", + "choice_c": "pad", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18495 + }, + { + "path": "1393010.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "horn", + "choice_d": "drummachine", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18496 + }, + { + "path": "1415796.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "piano", + "choice_b": "drummachine", + "choice_c": "harmonica", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18497 + }, + { + "path": "1076561.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "strings", + "choice_b": "accordion", + "choice_c": "saxophone", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18498 + }, + { + "path": "1163964.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "harmonica", + "choice_b": "acousticguitar", + "choice_c": "beat", + "choice_d": "sampler", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18499 + }, + { + "path": "1183238.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "piano", + "choice_b": "bass", + "choice_c": "acousticguitar", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18500 + }, + { + "path": "1292797.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "clarinet", + "choice_b": "classicalguitar", + "choice_c": "horn", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18501 + }, + { + "path": "1168675.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "electricpiano", + "choice_b": "piano", + "choice_c": "bell", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18502 + }, + { + "path": "896598.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "percussion", + "choice_b": "piano", + "choice_c": "trumpet", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18503 + }, + { + "path": "1339697.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "trumpet", + "choice_b": "pad", + "choice_c": "synthesizer", + "choice_d": "accordion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18504 + }, + { + "path": "1374846.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "piano", + "choice_b": "horn", + "choice_c": "pad", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18505 + }, + { + "path": "1389198.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "orchestra", + "choice_b": "flute", + "choice_c": "bass", + "choice_d": "cello", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18506 + }, + { + "path": "919996.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "piano", + "choice_b": "classicalguitar", + "choice_c": "guitar", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18507 + }, + { + "path": "1216038.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "guitar", + "choice_b": "drums", + "choice_c": "orchestra", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18508 + }, + { + "path": "80821.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "accordion", + "choice_b": "piano", + "choice_c": "pipeorgan", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18509 + }, + { + "path": "88556.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "classicalguitar", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18510 + }, + { + "path": "915288.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "synthesizer", + "choice_b": "drummachine", + "choice_c": "brass", + "choice_d": "voice", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18511 + }, + { + "path": "80823.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "electricguitar", + "choice_b": "beat", + "choice_c": "piano", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18512 + }, + { + "path": "361989.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "harp", + "choice_b": "keyboard", + "choice_c": "acousticguitar", + "choice_d": "bell", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18513 + }, + { + "path": "1227497.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "strings", + "choice_b": "classicalguitar", + "choice_c": "acousticguitar", + "choice_d": "percussion", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18514 + }, + { + "path": "147386.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "organ", + "choice_b": "beat", + "choice_c": "sampler", + "choice_d": "brass", + "answer_gt": "sampler", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18515 + }, + { + "path": "1102216.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "strings", + "choice_b": "computer", + "choice_c": "keyboard", + "choice_d": "piano", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18516 + }, + { + "path": "1174877.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "synthesizer", + "choice_b": "electricguitar", + "choice_c": "classicalguitar", + "choice_d": "trumpet", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18517 + }, + { + "path": "611731.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "beat", + "choice_d": "acousticguitar", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18518 + }, + { + "path": "1245890.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "drummachine", + "choice_b": "bass", + "choice_c": "cello", + "choice_d": "computer", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18519 + }, + { + "path": "259327.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "orchestra", + "choice_b": "classicalguitar", + "choice_c": "accordion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18520 + }, + { + "path": "1172395.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "piano", + "choice_b": "bell", + "choice_c": "harmonica", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18521 + }, + { + "path": "1286023.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "harmonica", + "choice_b": "synthesizer", + "choice_c": "bass", + "choice_d": "acousticguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18522 + }, + { + "path": "1133080.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "computer", + "choice_b": "pad", + "choice_c": "piano", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18523 + }, + { + "path": "1104648.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "clarinet", + "choice_b": "computer", + "choice_c": "drummachine", + "choice_d": "percussion", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18524 + }, + { + "path": "1395032.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "piano", + "choice_b": "orchestra", + "choice_c": "guitar", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18525 + }, + { + "path": "1077041.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "harp", + "choice_c": "piano", + "choice_d": "horn", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18526 + }, + { + "path": "511069.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "saxophone", + "choice_b": "guitar", + "choice_c": "bass", + "choice_d": "drums", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18527 + }, + { + "path": "14013.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "piano", + "choice_b": "harmonica", + "choice_c": "pipeorgan", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18528 + }, + { + "path": "1244713.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "trumpet", + "choice_b": "computer", + "choice_c": "synthesizer", + "choice_d": "strings", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18529 + }, + { + "path": "1349035.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "cello", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18530 + }, + { + "path": "1286022.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "synthesizer", + "choice_b": "trumpet", + "choice_c": "flute", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18531 + }, + { + "path": "589415.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "violin", + "choice_b": "electricguitar", + "choice_c": "strings", + "choice_d": "rhodes", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18532 + }, + { + "path": "1126844.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "pipeorgan", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18533 + }, + { + "path": "991647.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "saxophone", + "choice_b": "sampler", + "choice_c": "bass", + "choice_d": "electricpiano", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18534 + }, + { + "path": "1410638.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "rhodes", + "choice_b": "pad", + "choice_c": "drummachine", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18535 + }, + { + "path": "1133072.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "piano", + "choice_b": "orchestra", + "choice_c": "harp", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18536 + }, + { + "path": "1348859.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "trumpet", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "harp", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18537 + }, + { + "path": "1269565.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "drummachine", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18538 + }, + { + "path": "1217305.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "guitar", + "choice_b": "clarinet", + "choice_c": "drummachine", + "choice_d": "flute", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18539 + }, + { + "path": "82639.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "organ", + "choice_b": "drummachine", + "choice_c": "accordion", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18540 + }, + { + "path": "176090.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "violin", + "choice_b": "trumpet", + "choice_c": "saxophone", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18541 + }, + { + "path": "1276508.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "beat", + "choice_b": "percussion", + "choice_c": "harp", + "choice_d": "electricguitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18542 + }, + { + "path": "1415804.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "harmonica", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "sampler", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18543 + }, + { + "path": "1342578.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "clarinet", + "choice_b": "pipeorgan", + "choice_c": "percussion", + "choice_d": "violin", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18544 + }, + { + "path": "1085243.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "harp", + "choice_b": "organ", + "choice_c": "piano", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18545 + }, + { + "path": "1218785.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "classicalguitar", + "choice_b": "drummachine", + "choice_c": "saxophone", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18546 + }, + { + "path": "1068641.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "trumpet", + "choice_b": "computer", + "choice_c": "voice", + "choice_d": "sampler", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18547 + }, + { + "path": "1332222.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "brass", + "choice_b": "harmonica", + "choice_c": "percussion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18548 + }, + { + "path": "1168557.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "drummachine", + "choice_b": "cello", + "choice_c": "electricpiano", + "choice_d": "synthesizer", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18549 + }, + { + "path": "903597.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "piano", + "choice_b": "guitar", + "choice_c": "acousticguitar", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18550 + }, + { + "path": "1089084.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "acousticguitar", + "choice_b": "piano", + "choice_c": "bass", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18551 + }, + { + "path": "1133089.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "percussion", + "choice_c": "strings", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18552 + }, + { + "path": "1350855.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "synthesizer", + "choice_b": "piano", + "choice_c": "bass", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18553 + }, + { + "path": "1292338.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "electricguitar", + "choice_b": "violin", + "choice_c": "rhodes", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18554 + }, + { + "path": "1366003.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "cello", + "choice_b": "pad", + "choice_c": "guitar", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18555 + }, + { + "path": "1000085.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "beat", + "choice_b": "strings", + "choice_c": "piano", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18556 + }, + { + "path": "1056231.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "synthesizer", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18557 + }, + { + "path": "1086648.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "sampler", + "choice_b": "computer", + "choice_c": "violin", + "choice_d": "synthesizer", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18558 + }, + { + "path": "742096.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "acousticguitar", + "choice_b": "saxophone", + "choice_c": "sampler", + "choice_d": "rhodes", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18559 + }, + { + "path": "768512.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "drums", + "choice_b": "cello", + "choice_c": "bass", + "choice_d": "pipeorgan", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18560 + }, + { + "path": "265521.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "saxophone", + "choice_b": "electricguitar", + "choice_c": "computer", + "choice_d": "sampler", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18561 + }, + { + "path": "1222765.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "classicalguitar", + "choice_b": "synthesizer", + "choice_c": "harp", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18562 + }, + { + "path": "1374320.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "saxophone", + "choice_b": "electricpiano", + "choice_c": "electricguitar", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18563 + }, + { + "path": "1215952.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "beat", + "choice_b": "electricguitar", + "choice_c": "guitar", + "choice_d": "clarinet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18564 + }, + { + "path": "373353.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "bass", + "choice_b": "synthesizer", + "choice_c": "organ", + "choice_d": "trumpet", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18565 + }, + { + "path": "1159557.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "computer", + "choice_b": "electricguitar", + "choice_c": "classicalguitar", + "choice_d": "violin", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18566 + }, + { + "path": "25227.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "harmonica", + "choice_b": "electricguitar", + "choice_c": "beat", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18567 + }, + { + "path": "1196186.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "strings", + "choice_b": "percussion", + "choice_c": "harmonica", + "choice_d": "accordion", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18568 + }, + { + "path": "1053498.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "electricguitar", + "choice_b": "bell", + "choice_c": "synthesizer", + "choice_d": "accordion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18569 + }, + { + "path": "757497.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "electricguitar", + "choice_b": "piano", + "choice_c": "voice", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18570 + }, + { + "path": "191462.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "sampler", + "choice_b": "drummachine", + "choice_c": "synthesizer", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18571 + }, + { + "path": "1133085.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "strings", + "choice_b": "bell", + "choice_c": "piano", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18572 + }, + { + "path": "1386809.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "guitar", + "choice_b": "electricpiano", + "choice_c": "drums", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18573 + }, + { + "path": "595367.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "drummachine", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18574 + }, + { + "path": "1293742.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "harp", + "choice_b": "computer", + "choice_c": "sampler", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18575 + }, + { + "path": "80773.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "acousticguitar", + "choice_b": "piano", + "choice_c": "brass", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18576 + }, + { + "path": "80761.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "piano", + "choice_b": "computer", + "choice_c": "voice", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18577 + }, + { + "path": "1234533.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "rhodes", + "choice_c": "voice", + "choice_d": "cello", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18578 + }, + { + "path": "1321118.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "classicalguitar", + "choice_c": "trumpet", + "choice_d": "orchestra", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18579 + }, + { + "path": "1250859.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "accordion", + "choice_b": "synthesizer", + "choice_c": "horn", + "choice_d": "classicalguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18580 + }, + { + "path": "1085246.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "pad", + "choice_b": "piano", + "choice_c": "trumpet", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18581 + }, + { + "path": "14028.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "clarinet", + "choice_b": "piano", + "choice_c": "keyboard", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18582 + }, + { + "path": "1083530.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "rhodes", + "choice_b": "piano", + "choice_c": "strings", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18583 + }, + { + "path": "1371266.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "cello", + "choice_b": "saxophone", + "choice_c": "acousticguitar", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18584 + }, + { + "path": "76977.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "rhodes", + "choice_b": "guitar", + "choice_c": "cello", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18585 + }, + { + "path": "1133087.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "synthesizer", + "choice_b": "piano", + "choice_c": "harp", + "choice_d": "orchestra", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18586 + }, + { + "path": "782166.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "sampler", + "choice_b": "orchestra", + "choice_c": "computer", + "choice_d": "harmonica", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18587 + }, + { + "path": "1032646.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "drums", + "choice_b": "harmonica", + "choice_c": "cello", + "choice_d": "classicalguitar", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18588 + }, + { + "path": "1284782.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "keyboard", + "choice_d": "horn", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18589 + }, + { + "path": "1083536.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "bell", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18590 + }, + { + "path": "511078.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "harp", + "choice_b": "strings", + "choice_c": "guitar", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18591 + }, + { + "path": "80798.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "harp", + "choice_b": "piano", + "choice_c": "guitar", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18592 + }, + { + "path": "36701.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "pad", + "choice_b": "computer", + "choice_c": "brass", + "choice_d": "voice", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18593 + }, + { + "path": "1204622.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "bass", + "choice_b": "guitar", + "choice_c": "keyboard", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18594 + }, + { + "path": "1350988.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "piano", + "choice_b": "strings", + "choice_c": "percussion", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18595 + }, + { + "path": "1366863.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "guitar", + "choice_b": "sampler", + "choice_c": "accordion", + "choice_d": "trumpet", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18596 + }, + { + "path": "1243732.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "accordion", + "choice_b": "brass", + "choice_c": "trumpet", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18597 + }, + { + "path": "243736.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "computer", + "choice_b": "keyboard", + "choice_c": "electricguitar", + "choice_d": "voice", + "answer_gt": "voice", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18598 + }, + { + "path": "88558.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "voice", + "choice_b": "piano", + "choice_c": "accordion", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18599 + }, + { + "path": "1302283.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "voice", + "choice_b": "sampler", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18600 + }, + { + "path": "1185652.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "trumpet", + "choice_b": "piano", + "choice_c": "beat", + "choice_d": "accordion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18601 + }, + { + "path": "1348861.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "voice", + "choice_b": "synthesizer", + "choice_c": "electricguitar", + "choice_d": "keyboard", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18602 + }, + { + "path": "1086645.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "percussion", + "choice_b": "electricpiano", + "choice_c": "violin", + "choice_d": "strings", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18603 + }, + { + "path": "1204591.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "drummachine", + "choice_b": "acousticguitar", + "choice_c": "horn", + "choice_d": "cello", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18604 + }, + { + "path": "1219427.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "electricpiano", + "choice_b": "computer", + "choice_c": "synthesizer", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18605 + }, + { + "path": "38122.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "flute", + "choice_b": "synthesizer", + "choice_c": "organ", + "choice_d": "pad", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18606 + }, + { + "path": "1149753.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "piano", + "choice_b": "organ", + "choice_c": "pad", + "choice_d": "trumpet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18607 + }, + { + "path": "1223053.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "sampler", + "choice_c": "flute", + "choice_d": "keyboard", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18608 + }, + { + "path": "80757.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "guitar", + "choice_b": "clarinet", + "choice_c": "piano", + "choice_d": "organ", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18609 + }, + { + "path": "979839.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "sampler", + "choice_c": "voice", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18610 + }, + { + "path": "941854.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "computer", + "choice_c": "classicalguitar", + "choice_d": "electricpiano", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18611 + }, + { + "path": "516732.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "drums", + "choice_b": "synthesizer", + "choice_c": "electricguitar", + "choice_d": "bell", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18612 + }, + { + "path": "1126845.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "violin", + "choice_b": "computer", + "choice_c": "strings", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18613 + }, + { + "path": "1202100.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "flute", + "choice_b": "strings", + "choice_c": "bass", + "choice_d": "acousticguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18614 + }, + { + "path": "1258187.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "beat", + "choice_b": "trumpet", + "choice_c": "drums", + "choice_d": "electricguitar", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18615 + }, + { + "path": "1244712.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "synthesizer", + "choice_b": "keyboard", + "choice_c": "saxophone", + "choice_d": "electricguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18616 + }, + { + "path": "73073.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "flute", + "choice_b": "synthesizer", + "choice_c": "drummachine", + "choice_d": "bass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18617 + }, + { + "path": "80791.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "piano", + "choice_b": "strings", + "choice_c": "horn", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18618 + }, + { + "path": "1319499.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "flute", + "choice_b": "horn", + "choice_c": "electricguitar", + "choice_d": "organ", + "answer_gt": "horn", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18619 + }, + { + "path": "1314148.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "drums", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "violin", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18620 + }, + { + "path": "1209206.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "computer", + "choice_b": "classicalguitar", + "choice_c": "rhodes", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18621 + }, + { + "path": "1417803.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "saxophone", + "choice_b": "accordion", + "choice_c": "percussion", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18622 + }, + { + "path": "1312110.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "beat", + "choice_b": "strings", + "choice_c": "cello", + "choice_d": "keyboard", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18623 + }, + { + "path": "1401014.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "sampler", + "choice_b": "acousticguitar", + "choice_c": "brass", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18624 + }, + { + "path": "1077727.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "synthesizer", + "choice_b": "harp", + "choice_c": "beat", + "choice_d": "flute", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18625 + }, + { + "path": "1416195.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "computer", + "choice_b": "synthesizer", + "choice_c": "harmonica", + "choice_d": "keyboard", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18626 + }, + { + "path": "979872.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "clarinet", + "choice_b": "electricpiano", + "choice_c": "sampler", + "choice_d": "trumpet", + "answer_gt": "sampler", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18627 + }, + { + "path": "1348655.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "horn", + "choice_b": "drummachine", + "choice_c": "saxophone", + "choice_d": "bass", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18628 + }, + { + "path": "82640.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "voice", + "choice_b": "cello", + "choice_c": "synthesizer", + "choice_d": "bass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18629 + }, + { + "path": "1314124.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "pipeorgan", + "choice_b": "pad", + "choice_c": "strings", + "choice_d": "acousticguitar", + "answer_gt": "pad", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18630 + }, + { + "path": "1288421.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "voice", + "choice_b": "pipeorgan", + "choice_c": "percussion", + "choice_d": "violin", + "answer_gt": "percussion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18631 + }, + { + "path": "80760.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "cello", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18632 + }, + { + "path": "1083525.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "organ", + "choice_b": "drums", + "choice_c": "strings", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18633 + }, + { + "path": "1394752.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "flute", + "choice_b": "harmonica", + "choice_c": "rhodes", + "choice_d": "electricguitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18634 + }, + { + "path": "747693.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "electricguitar", + "choice_b": "cello", + "choice_c": "synthesizer", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18635 + }, + { + "path": "1168542.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "voice", + "choice_b": "electricpiano", + "choice_c": "percussion", + "choice_d": "keyboard", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18636 + }, + { + "path": "1047958.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "synthesizer", + "choice_b": "clarinet", + "choice_c": "drummachine", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18637 + }, + { + "path": "1044925.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "pad", + "choice_b": "computer", + "choice_c": "orchestra", + "choice_d": "keyboard", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18638 + }, + { + "path": "1078368.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "clarinet", + "choice_b": "harp", + "choice_c": "horn", + "choice_d": "strings", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18639 + }, + { + "path": "1250857.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "drums", + "choice_b": "pipeorgan", + "choice_c": "orchestra", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18640 + }, + { + "path": "1312106.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "trumpet", + "choice_b": "clarinet", + "choice_c": "violin", + "choice_d": "bass", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18641 + }, + { + "path": "175423.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "flute", + "choice_b": "keyboard", + "choice_c": "pad", + "choice_d": "harp", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18642 + }, + { + "path": "1266724.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "trumpet", + "choice_b": "saxophone", + "choice_c": "synthesizer", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18643 + }, + { + "path": "1250862.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "electricpiano", + "choice_b": "synthesizer", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18644 + }, + { + "path": "1011995.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "harp", + "choice_b": "piano", + "choice_c": "bell", + "choice_d": "harmonica", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18645 + }, + { + "path": "875373.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "saxophone", + "choice_b": "horn", + "choice_c": "synthesizer", + "choice_d": "brass", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18646 + }, + { + "path": "23861.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "acousticguitar", + "choice_b": "saxophone", + "choice_c": "keyboard", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18647 + }, + { + "path": "1168545.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "organ", + "choice_b": "harp", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18648 + }, + { + "path": "25219.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "computer", + "choice_b": "trumpet", + "choice_c": "orchestra", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18649 + }, + { + "path": "1086829.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "violin", + "choice_b": "trumpet", + "choice_c": "pipeorgan", + "choice_d": "keyboard", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18650 + }, + { + "path": "742058.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "strings", + "choice_b": "accordion", + "choice_c": "classicalguitar", + "choice_d": "bell", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18651 + }, + { + "path": "1408511.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "orchestra", + "choice_b": "guitar", + "choice_c": "brass", + "choice_d": "harmonica", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18652 + }, + { + "path": "1047579.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "rhodes", + "choice_b": "bass", + "choice_c": "acousticguitar", + "choice_d": "brass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18653 + }, + { + "path": "828028.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "harmonica", + "choice_b": "harp", + "choice_c": "synthesizer", + "choice_d": "bell", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18654 + }, + { + "path": "965752.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "drummachine", + "choice_b": "bell", + "choice_c": "guitar", + "choice_d": "electricguitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18655 + }, + { + "path": "398182.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "flute", + "choice_b": "piano", + "choice_c": "rhodes", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18656 + }, + { + "path": "80780.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "guitar", + "choice_b": "flute", + "choice_c": "organ", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18657 + }, + { + "path": "80776.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "drummachine", + "choice_b": "voice", + "choice_c": "trumpet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18658 + }, + { + "path": "1076581.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "cello", + "choice_b": "rhodes", + "choice_c": "drums", + "choice_d": "beat", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18659 + }, + { + "path": "1418655.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "harmonica", + "choice_b": "trumpet", + "choice_c": "guitar", + "choice_d": "bell", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18660 + }, + { + "path": "1243723.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "beat", + "choice_d": "percussion", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18661 + }, + { + "path": "1078363.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "drummachine", + "choice_b": "electricguitar", + "choice_c": "harp", + "choice_d": "harmonica", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18662 + }, + { + "path": "736793.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "strings", + "choice_b": "sampler", + "choice_c": "harp", + "choice_d": "cello", + "answer_gt": "cello", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18663 + }, + { + "path": "1164338.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "strings", + "choice_b": "horn", + "choice_c": "synthesizer", + "choice_d": "clarinet", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18664 + }, + { + "path": "945025.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "rhodes", + "choice_b": "orchestra", + "choice_c": "beat", + "choice_d": "electricpiano", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18665 + }, + { + "path": "833459.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "piano", + "choice_b": "computer", + "choice_c": "guitar", + "choice_d": "orchestra", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18666 + }, + { + "path": "1379936.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "pad", + "choice_b": "voice", + "choice_c": "piano", + "choice_d": "sampler", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18667 + }, + { + "path": "1105199.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "drummachine", + "choice_c": "organ", + "choice_d": "orchestra", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18668 + }, + { + "path": "1109581.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "beat", + "choice_b": "flute", + "choice_c": "pad", + "choice_d": "accordion", + "answer_gt": "accordion", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18669 + }, + { + "path": "1210604.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "piano", + "choice_b": "horn", + "choice_c": "drummachine", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18670 + }, + { + "path": "1354436.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "voice", + "choice_b": "drums", + "choice_c": "pipeorgan", + "choice_d": "harmonica", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18671 + }, + { + "path": "1374335.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "electricguitar", + "choice_b": "classicalguitar", + "choice_c": "beat", + "choice_d": "flute", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18672 + }, + { + "path": "1400502.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "trumpet", + "choice_b": "acousticguitar", + "choice_c": "piano", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18673 + }, + { + "path": "1164047.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "bell", + "choice_b": "drums", + "choice_c": "electricpiano", + "choice_d": "acousticguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18674 + }, + { + "path": "1273388.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "guitar", + "choice_b": "bass", + "choice_c": "trumpet", + "choice_d": "pad", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18675 + }, + { + "path": "1053504.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "flute", + "choice_b": "electricpiano", + "choice_c": "synthesizer", + "choice_d": "harmonica", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18676 + }, + { + "path": "1083511.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "saxophone", + "choice_b": "classicalguitar", + "choice_c": "pad", + "choice_d": "bass", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18677 + }, + { + "path": "511086.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "pipeorgan", + "choice_b": "guitar", + "choice_c": "organ", + "choice_d": "saxophone", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18678 + }, + { + "path": "1164366.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "beat", + "choice_b": "drummachine", + "choice_c": "bell", + "choice_d": "keyboard", + "answer_gt": "bell", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18679 + }, + { + "path": "1086649.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "brass", + "choice_b": "harp", + "choice_c": "violin", + "choice_d": "percussion", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18680 + }, + { + "path": "80820.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "percussion", + "choice_b": "piano", + "choice_c": "accordion", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18681 + }, + { + "path": "184197.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "voice", + "choice_b": "saxophone", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18682 + }, + { + "path": "1306074.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "violin", + "choice_b": "computer", + "choice_c": "brass", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18683 + }, + { + "path": "752315.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "piano", + "choice_c": "acousticguitar", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18684 + }, + { + "path": "1174882.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "classicalguitar", + "choice_b": "pipeorgan", + "choice_c": "drums", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18685 + }, + { + "path": "1386750.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "clarinet", + "choice_b": "beat", + "choice_c": "guitar", + "choice_d": "rhodes", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18686 + }, + { + "path": "80817.mp3", + "question": "What instrument is at the forefront of this musical work?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "computer", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18687 + }, + { + "path": "1266723.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "electricguitar", + "choice_b": "percussion", + "choice_c": "harmonica", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18688 + }, + { + "path": "277804.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "computer", + "choice_b": "cello", + "choice_c": "acousticguitar", + "choice_d": "classicalguitar", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18689 + }, + { + "path": "1085241.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "harmonica", + "choice_b": "acousticguitar", + "choice_c": "trumpet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18690 + }, + { + "path": "967565.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "electricpiano", + "choice_b": "horn", + "choice_c": "clarinet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18691 + }, + { + "path": "361994.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "acousticguitar", + "choice_b": "electricpiano", + "choice_c": "bell", + "choice_d": "flute", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18692 + }, + { + "path": "1133084.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "piano", + "choice_b": "acousticguitar", + "choice_c": "harp", + "choice_d": "synthesizer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18693 + }, + { + "path": "1371041.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "synthesizer", + "choice_b": "accordion", + "choice_c": "beat", + "choice_d": "pipeorgan", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18694 + }, + { + "path": "1062501.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "drums", + "choice_b": "synthesizer", + "choice_c": "beat", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18695 + }, + { + "path": "397787.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "guitar", + "choice_b": "brass", + "choice_c": "harmonica", + "choice_d": "accordion", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18696 + }, + { + "path": "9600.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "clarinet", + "choice_b": "voice", + "choice_c": "acousticguitar", + "choice_d": "pipeorgan", + "answer_gt": "pipeorgan", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18697 + }, + { + "path": "516737.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "harp", + "choice_b": "pad", + "choice_c": "synthesizer", + "choice_d": "organ", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18698 + }, + { + "path": "1119921.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "keyboard", + "choice_b": "accordion", + "choice_c": "harmonica", + "choice_d": "classicalguitar", + "answer_gt": "harmonica", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18699 + }, + { + "path": "1372208.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "guitar", + "choice_b": "harp", + "choice_c": "piano", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18700 + }, + { + "path": "915293.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "drummachine", + "choice_b": "guitar", + "choice_c": "sampler", + "choice_d": "saxophone", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18701 + }, + { + "path": "1244465.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "brass", + "choice_b": "synthesizer", + "choice_c": "bass", + "choice_d": "trumpet", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18702 + }, + { + "path": "1355739.mp3", + "question": "Which instrument is the highlight of this composition?", + "choice_a": "trumpet", + "choice_b": "pipeorgan", + "choice_c": "horn", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18703 + }, + { + "path": "1122856.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "harmonica", + "choice_b": "saxophone", + "choice_c": "synthesizer", + "choice_d": "drummachine", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18704 + }, + { + "path": "1403977.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "piano", + "choice_b": "accordion", + "choice_c": "brass", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18705 + }, + { + "path": "1163779.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "synthesizer", + "choice_b": "cello", + "choice_c": "trumpet", + "choice_d": "violin", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18706 + }, + { + "path": "1227495.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "orchestra", + "choice_b": "flute", + "choice_c": "saxophone", + "choice_d": "brass", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18707 + }, + { + "path": "36704.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "clarinet", + "choice_b": "synthesizer", + "choice_c": "flute", + "choice_d": "computer", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18708 + }, + { + "path": "1117577.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "electricguitar", + "choice_b": "pad", + "choice_c": "organ", + "choice_d": "acousticguitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18709 + }, + { + "path": "925507.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "voice", + "choice_b": "electricguitar", + "choice_c": "piano", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18710 + }, + { + "path": "816198.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "voice", + "choice_b": "piano", + "choice_c": "orchestra", + "choice_d": "bass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18711 + }, + { + "path": "1268996.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "saxophone", + "choice_b": "harp", + "choice_c": "guitar", + "choice_d": "electricpiano", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18712 + }, + { + "path": "977854.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "electricpiano", + "choice_b": "drums", + "choice_c": "sampler", + "choice_d": "organ", + "answer_gt": "electricpiano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18713 + }, + { + "path": "76972.mp3", + "question": "Which instrument takes center stage in this music?", + "choice_a": "clarinet", + "choice_b": "orchestra", + "choice_c": "classicalguitar", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18714 + }, + { + "path": "111584.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "horn", + "choice_b": "brass", + "choice_c": "drums", + "choice_d": "synthesizer", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18715 + }, + { + "path": "1332220.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "drums", + "choice_b": "piano", + "choice_c": "electricpiano", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18716 + }, + { + "path": "979851.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "drums", + "choice_b": "drummachine", + "choice_c": "synthesizer", + "choice_d": "sampler", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18717 + }, + { + "path": "171917.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "horn", + "choice_b": "synthesizer", + "choice_c": "voice", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18718 + }, + { + "path": "1053497.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "violin", + "choice_b": "synthesizer", + "choice_c": "orchestra", + "choice_d": "classicalguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18719 + }, + { + "path": "1079253.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "piano", + "choice_b": "saxophone", + "choice_c": "drummachine", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18720 + }, + { + "path": "1197780.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "accordion", + "choice_b": "guitar", + "choice_c": "flute", + "choice_d": "beat", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18721 + }, + { + "path": "1284439.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "keyboard", + "choice_b": "trumpet", + "choice_c": "pad", + "choice_d": "rhodes", + "answer_gt": "keyboard", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18722 + }, + { + "path": "1342576.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "strings", + "choice_b": "sampler", + "choice_c": "saxophone", + "choice_d": "harp", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18723 + }, + { + "path": "1085248.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "violin", + "choice_b": "electricpiano", + "choice_c": "flute", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18724 + }, + { + "path": "1374316.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "pad", + "choice_b": "cello", + "choice_c": "pipeorgan", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18725 + }, + { + "path": "149490.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "sampler", + "choice_b": "clarinet", + "choice_c": "piano", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18726 + }, + { + "path": "1121565.mp3", + "question": "What's the chief instrument you hear in this piece of music?", + "choice_a": "guitar", + "choice_b": "acousticguitar", + "choice_c": "electricguitar", + "choice_d": "brass", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18727 + }, + { + "path": "921508.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bass", + "choice_b": "pipeorgan", + "choice_c": "saxophone", + "choice_d": "sampler", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18728 + }, + { + "path": "1160224.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "classicalguitar", + "choice_b": "drums", + "choice_c": "trumpet", + "choice_d": "rhodes", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18729 + }, + { + "path": "1086828.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "accordion", + "choice_b": "violin", + "choice_c": "guitar", + "choice_d": "drums", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18730 + }, + { + "path": "41967.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "voice", + "choice_b": "beat", + "choice_c": "harmonica", + "choice_d": "percussion", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18731 + }, + { + "path": "25226.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "brass", + "choice_b": "piano", + "choice_c": "bass", + "choice_d": "horn", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18732 + }, + { + "path": "1053499.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "synthesizer", + "choice_b": "harmonica", + "choice_c": "guitar", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18733 + }, + { + "path": "159622.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "synthesizer", + "choice_b": "guitar", + "choice_c": "bell", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18734 + }, + { + "path": "1418665.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "bass", + "choice_b": "synthesizer", + "choice_c": "organ", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18735 + }, + { + "path": "914769.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "acousticguitar", + "choice_b": "synthesizer", + "choice_c": "orchestra", + "choice_d": "bass", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18736 + }, + { + "path": "914773.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "guitar", + "choice_b": "harp", + "choice_c": "acousticguitar", + "choice_d": "piano", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18737 + }, + { + "path": "1306024.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "bell", + "choice_b": "drums", + "choice_c": "saxophone", + "choice_d": "drummachine", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18738 + }, + { + "path": "1346878.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "drummachine", + "choice_b": "strings", + "choice_c": "rhodes", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18739 + }, + { + "path": "1297484.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "piano", + "choice_b": "electricguitar", + "choice_c": "bass", + "choice_d": "cello", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18740 + }, + { + "path": "1060966.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "harmonica", + "choice_b": "pad", + "choice_c": "acousticguitar", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18741 + }, + { + "path": "1205809.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "strings", + "choice_d": "flute", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18742 + }, + { + "path": "1275967.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "bass", + "choice_b": "sampler", + "choice_c": "strings", + "choice_d": "keyboard", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18743 + }, + { + "path": "1268990.mp3", + "question": "determine the instrument that's most pronounced in this track.", + "choice_a": "pad", + "choice_b": "guitar", + "choice_c": "classicalguitar", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18744 + }, + { + "path": "1073702.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "strings", + "choice_b": "bass", + "choice_c": "bell", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18745 + }, + { + "path": "1257052.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "electricpiano", + "choice_b": "electricguitar", + "choice_c": "piano", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18746 + }, + { + "path": "1376467.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "rhodes", + "choice_b": "guitar", + "choice_c": "harmonica", + "choice_d": "orchestra", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18747 + }, + { + "path": "1083529.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "clarinet", + "choice_b": "guitar", + "choice_c": "beat", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18748 + }, + { + "path": "14024.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "piano", + "choice_b": "harmonica", + "choice_c": "flute", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18749 + }, + { + "path": "1276513.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "beat", + "choice_b": "violin", + "choice_c": "electricguitar", + "choice_d": "drummachine", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18750 + }, + { + "path": "335254.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "harmonica", + "choice_b": "beat", + "choice_c": "piano", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18751 + }, + { + "path": "575456.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "pipeorgan", + "choice_b": "trumpet", + "choice_c": "acousticguitar", + "choice_d": "strings", + "answer_gt": "acousticguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18752 + }, + { + "path": "1297481.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "pipeorgan", + "choice_b": "computer", + "choice_c": "bass", + "choice_d": "organ", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18753 + }, + { + "path": "1196187.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "accordion", + "choice_b": "strings", + "choice_c": "computer", + "choice_d": "rhodes", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18754 + }, + { + "path": "1330560.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "guitar", + "choice_b": "synthesizer", + "choice_c": "acousticguitar", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18755 + }, + { + "path": "1389206.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "harp", + "choice_b": "orchestra", + "choice_c": "synthesizer", + "choice_d": "accordion", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18756 + }, + { + "path": "1243736.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "beat", + "choice_b": "voice", + "choice_c": "clarinet", + "choice_d": "classicalguitar", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18757 + }, + { + "path": "1204669.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "trumpet", + "choice_b": "classicalguitar", + "choice_c": "guitar", + "choice_d": "pipeorgan", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18758 + }, + { + "path": "414045.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "strings", + "choice_c": "piano", + "choice_d": "cello", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18759 + }, + { + "path": "1214873.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "electricpiano", + "choice_b": "accordion", + "choice_c": "synthesizer", + "choice_d": "clarinet", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18760 + }, + { + "path": "511074.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "guitar", + "choice_b": "strings", + "choice_c": "electricpiano", + "choice_d": "organ", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18761 + }, + { + "path": "757496.mp3", + "question": "What instrument is the linchpin of this musical piece?", + "choice_a": "piano", + "choice_b": "flute", + "choice_c": "bell", + "choice_d": "pad", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18762 + }, + { + "path": "1203896.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "bell", + "choice_b": "drums", + "choice_c": "piano", + "choice_d": "trumpet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18763 + }, + { + "path": "1131997.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "accordion", + "choice_b": "pad", + "choice_c": "orchestra", + "choice_d": "sampler", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18764 + }, + { + "path": "1044930.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "computer", + "choice_b": "orchestra", + "choice_c": "piano", + "choice_d": "electricguitar", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18765 + }, + { + "path": "802421.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "electricguitar", + "choice_b": "acousticguitar", + "choice_c": "bass", + "choice_d": "guitar", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18766 + }, + { + "path": "1062506.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "harmonica", + "choice_b": "synthesizer", + "choice_c": "bell", + "choice_d": "drums", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18767 + }, + { + "path": "30225.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "voice", + "choice_b": "guitar", + "choice_c": "classicalguitar", + "choice_d": "brass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18768 + }, + { + "path": "1086429.mp3", + "question": "What's the most critical instrument in this music piece?", + "choice_a": "violin", + "choice_b": "sampler", + "choice_c": "accordion", + "choice_d": "classicalguitar", + "answer_gt": "violin", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18769 + }, + { + "path": "373056.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "flute", + "choice_b": "electricguitar", + "choice_c": "piano", + "choice_d": "harmonica", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18770 + }, + { + "path": "1280191.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "organ", + "choice_b": "guitar", + "choice_c": "orchestra", + "choice_d": "synthesizer", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18771 + }, + { + "path": "903590.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "electricpiano", + "choice_b": "piano", + "choice_c": "beat", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18772 + }, + { + "path": "1026344.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "percussion", + "choice_b": "violin", + "choice_c": "accordion", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18773 + }, + { + "path": "1228696.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "piano", + "choice_b": "rhodes", + "choice_c": "orchestra", + "choice_d": "bell", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18774 + }, + { + "path": "1380732.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "bass", + "choice_b": "percussion", + "choice_c": "strings", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18775 + }, + { + "path": "1115567.mp3", + "question": "name the instrument that's at the helm of this track.", + "choice_a": "percussion", + "choice_b": "guitar", + "choice_c": "computer", + "choice_d": "acousticguitar", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18776 + }, + { + "path": "1077721.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "sampler", + "choice_b": "drums", + "choice_c": "drummachine", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18777 + }, + { + "path": "1116712.mp3", + "question": "What instrument is most prominent in this piece?", + "choice_a": "drums", + "choice_b": "classicalguitar", + "choice_c": "flute", + "choice_d": "drummachine", + "answer_gt": "classicalguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18778 + }, + { + "path": "1389037.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "pipeorgan", + "choice_b": "drummachine", + "choice_c": "guitar", + "choice_d": "drums", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18779 + }, + { + "path": "362637.mp3", + "question": "What's the leading instrument featured in this track?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "classicalguitar", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18780 + }, + { + "path": "40630.mp3", + "question": "What's the most noticeable instrument in this composition?", + "choice_a": "piano", + "choice_b": "percussion", + "choice_c": "drums", + "choice_d": "beat", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18781 + }, + { + "path": "915294.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "drummachine", + "choice_b": "percussion", + "choice_c": "piano", + "choice_d": "bell", + "answer_gt": "drummachine", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18782 + }, + { + "path": "1041289.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "flute", + "choice_b": "brass", + "choice_c": "cello", + "choice_d": "strings", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18783 + }, + { + "path": "1314127.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "bell", + "choice_b": "violin", + "choice_c": "percussion", + "choice_d": "acousticguitar", + "answer_gt": "bell", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18784 + }, + { + "path": "1041284.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "guitar", + "choice_b": "piano", + "choice_c": "flute", + "choice_d": "electricpiano", + "answer_gt": "flute", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18785 + }, + { + "path": "1203655.mp3", + "question": "Which instrument takes the lead role in this composition?", + "choice_a": "piano", + "choice_b": "saxophone", + "choice_c": "brass", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18786 + }, + { + "path": "1104646.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "pipeorgan", + "choice_b": "computer", + "choice_c": "electricpiano", + "choice_d": "brass", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18787 + }, + { + "path": "1117566.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "pad", + "choice_b": "orchestra", + "choice_c": "trumpet", + "choice_d": "electricpiano", + "answer_gt": "trumpet", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18788 + }, + { + "path": "1326462.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "orchestra", + "choice_b": "synthesizer", + "choice_c": "classicalguitar", + "choice_d": "trumpet", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18789 + }, + { + "path": "1196184.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "violin", + "choice_b": "strings", + "choice_c": "electricpiano", + "choice_d": "pad", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18790 + }, + { + "path": "742093.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "pad", + "choice_b": "horn", + "choice_c": "harp", + "choice_d": "saxophone", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18791 + }, + { + "path": "1278515.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "electricpiano", + "choice_b": "orchestra", + "choice_c": "clarinet", + "choice_d": "violin", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18792 + }, + { + "path": "73080.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "harp", + "choice_b": "horn", + "choice_c": "computer", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18793 + }, + { + "path": "1133076.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "bass", + "choice_b": "violin", + "choice_c": "piano", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18794 + }, + { + "path": "1217980.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "accordion", + "choice_c": "rhodes", + "choice_d": "percussion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18795 + }, + { + "path": "1346883.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "drummachine", + "choice_b": "synthesizer", + "choice_c": "trumpet", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18796 + }, + { + "path": "80756.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "orchestra", + "choice_b": "piano", + "choice_c": "trumpet", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18797 + }, + { + "path": "88561.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "piano", + "choice_b": "bell", + "choice_c": "keyboard", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18798 + }, + { + "path": "881054.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "synthesizer", + "choice_b": "bell", + "choice_c": "violin", + "choice_d": "strings", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18799 + }, + { + "path": "173545.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "drums", + "choice_b": "clarinet", + "choice_c": "guitar", + "choice_d": "sampler", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18800 + }, + { + "path": "1267752.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "rhodes", + "choice_b": "beat", + "choice_c": "piano", + "choice_d": "drummachine", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18801 + }, + { + "path": "1196185.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bell", + "choice_b": "strings", + "choice_c": "piano", + "choice_d": "guitar", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18802 + }, + { + "path": "80800.mp3", + "question": "Which instrument is the star of this musical piece?", + "choice_a": "flute", + "choice_b": "piano", + "choice_c": "brass", + "choice_d": "strings", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18803 + }, + { + "path": "940188.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "beat", + "choice_b": "strings", + "choice_c": "rhodes", + "choice_d": "pipeorgan", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18804 + }, + { + "path": "1159540.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "beat", + "choice_b": "percussion", + "choice_c": "piano", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18805 + }, + { + "path": "23862.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18806 + }, + { + "path": "859972.mp3", + "question": "What's the most significant instrument in this music piece?", + "choice_a": "horn", + "choice_b": "synthesizer", + "choice_c": "violin", + "choice_d": "electricpiano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18807 + }, + { + "path": "1397348.mp3", + "question": "Which instrument is the driving force behind this piece?", + "choice_a": "trumpet", + "choice_b": "bass", + "choice_c": "piano", + "choice_d": "keyboard", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18808 + }, + { + "path": "80812.mp3", + "question": "specify the instrument that's most featured in this track.", + "choice_a": "bell", + "choice_b": "bass", + "choice_c": "percussion", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18809 + }, + { + "path": "1291475.mp3", + "question": "What's the major instrument you can hear in this music?", + "choice_a": "brass", + "choice_b": "guitar", + "choice_c": "percussion", + "choice_d": "harp", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18810 + }, + { + "path": "1076566.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "guitar", + "choice_b": "beat", + "choice_c": "rhodes", + "choice_d": "classicalguitar", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18811 + }, + { + "path": "1300027.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "brass", + "choice_b": "piano", + "choice_c": "electricpiano", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18812 + }, + { + "path": "875378.mp3", + "question": "What instrument is the heart of this musical track?", + "choice_a": "electricpiano", + "choice_b": "brass", + "choice_c": "synthesizer", + "choice_d": "rhodes", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18813 + }, + { + "path": "23866.mp3", + "question": "What is the main instrument used to play this piece of music?", + "choice_a": "horn", + "choice_b": "drums", + "choice_c": "trumpet", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18814 + }, + { + "path": "80759.mp3", + "question": "pinpoint the instrument that stands out in this track.", + "choice_a": "piano", + "choice_b": "classicalguitar", + "choice_c": "acousticguitar", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18815 + }, + { + "path": "1245729.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "voice", + "choice_c": "harp", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18816 + }, + { + "path": "1291477.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "acousticguitar", + "choice_b": "bass", + "choice_c": "harmonica", + "choice_d": "flute", + "answer_gt": "harmonica", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18817 + }, + { + "path": "736523.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "voice", + "choice_b": "brass", + "choice_c": "violin", + "choice_d": "bell", + "answer_gt": "brass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18818 + }, + { + "path": "17177.mp3", + "question": "What's the instrumental focus of this piece of music?", + "choice_a": "rhodes", + "choice_b": "saxophone", + "choice_c": "orchestra", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18819 + }, + { + "path": "88553.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "sampler", + "choice_b": "electricguitar", + "choice_c": "pipeorgan", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18820 + }, + { + "path": "1133088.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "horn", + "choice_b": "accordion", + "choice_c": "cello", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18821 + }, + { + "path": "287786.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "synthesizer", + "choice_b": "clarinet", + "choice_c": "orchestra", + "choice_d": "percussion", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18822 + }, + { + "path": "57266.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "clarinet", + "choice_b": "violin", + "choice_c": "saxophone", + "choice_d": "bass", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18823 + }, + { + "path": "1216967.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "drums", + "choice_b": "computer", + "choice_c": "guitar", + "choice_d": "piano", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18824 + }, + { + "path": "1244716.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "strings", + "choice_b": "brass", + "choice_c": "synthesizer", + "choice_d": "acousticguitar", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18825 + }, + { + "path": "736787.mp3", + "question": "detect the main instrument that defines this track.", + "choice_a": "guitar", + "choice_b": "organ", + "choice_c": "cello", + "choice_d": "bell", + "answer_gt": "cello", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18826 + }, + { + "path": "88554.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "piano", + "choice_b": "trumpet", + "choice_c": "saxophone", + "choice_d": "drums", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18827 + }, + { + "path": "1347935.mp3", + "question": "Which instrument is the pillar of this composition?", + "choice_a": "keyboard", + "choice_b": "acousticguitar", + "choice_c": "synthesizer", + "choice_d": "flute", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18828 + }, + { + "path": "80797.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "horn", + "choice_b": "strings", + "choice_c": "piano", + "choice_d": "violin", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18829 + }, + { + "path": "1354790.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "rhodes", + "choice_b": "flute", + "choice_c": "harp", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18830 + }, + { + "path": "1356518.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "synthesizer", + "choice_b": "sampler", + "choice_c": "piano", + "choice_d": "cello", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18831 + }, + { + "path": "757499.mp3", + "question": "What's the primary source of sound in this music track?", + "choice_a": "beat", + "choice_b": "piano", + "choice_c": "harp", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18832 + }, + { + "path": "14021.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "piano", + "choice_b": "classicalguitar", + "choice_c": "pad", + "choice_d": "clarinet", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18833 + }, + { + "path": "131636.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "guitar", + "choice_b": "keyboard", + "choice_c": "harmonica", + "choice_d": "pipeorgan", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18834 + }, + { + "path": "244956.mp3", + "question": "What's the key instrument played in this music piece?", + "choice_a": "flute", + "choice_b": "piano", + "choice_c": "voice", + "choice_d": "acousticguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18835 + }, + { + "path": "575216.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "guitar", + "choice_b": "drummachine", + "choice_c": "piano", + "choice_d": "drums", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18836 + }, + { + "path": "1196183.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "bell", + "choice_b": "strings", + "choice_c": "acousticguitar", + "choice_d": "organ", + "answer_gt": "strings", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18837 + }, + { + "path": "1319330.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "rhodes", + "choice_b": "accordion", + "choice_c": "synthesizer", + "choice_d": "saxophone", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18838 + }, + { + "path": "398184.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "piano", + "choice_b": "orchestra", + "choice_c": "drums", + "choice_d": "harp", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18839 + }, + { + "path": "1047960.mp3", + "question": "Which instrument is the bedrock of this composition?", + "choice_a": "rhodes", + "choice_b": "synthesizer", + "choice_c": "harp", + "choice_d": "beat", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18840 + }, + { + "path": "1350823.mp3", + "question": "What's the central instrument in this piece of music?", + "choice_a": "cello", + "choice_b": "guitar", + "choice_c": "rhodes", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18841 + }, + { + "path": "1131998.mp3", + "question": "What instrument is primarily responsible for the melody in this track?", + "choice_a": "drums", + "choice_b": "orchestra", + "choice_c": "harmonica", + "choice_d": "trumpet", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18842 + }, + { + "path": "1419768.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "piano", + "choice_b": "pad", + "choice_c": "sampler", + "choice_d": "brass", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18843 + }, + { + "path": "77442.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "sampler", + "choice_b": "drums", + "choice_c": "horn", + "choice_d": "strings", + "answer_gt": "drums", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18844 + }, + { + "path": "1084130.mp3", + "question": "What instrument is the essence of this musical track?", + "choice_a": "piano", + "choice_b": "sampler", + "choice_c": "computer", + "choice_d": "trumpet", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18845 + }, + { + "path": "1297477.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "rhodes", + "choice_b": "pad", + "choice_c": "bass", + "choice_d": "harp", + "answer_gt": "bass", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18846 + }, + { + "path": "1023481.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "electricguitar", + "choice_b": "piano", + "choice_c": "pipeorgan", + "choice_d": "flute", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18847 + }, + { + "path": "1214874.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "synthesizer", + "choice_b": "brass", + "choice_c": "flute", + "choice_d": "horn", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18848 + }, + { + "path": "847261.mp3", + "question": "tell me the dominant instrument in this composition?", + "choice_a": "piano", + "choice_b": "flute", + "choice_c": "electricguitar", + "choice_d": "computer", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18849 + }, + { + "path": "80793.mp3", + "question": "What's the foremost instrument used in this musical number?", + "choice_a": "piano", + "choice_b": "rhodes", + "choice_c": "pad", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18850 + }, + { + "path": "131600.mp3", + "question": "identify the primary instrument in this musical piece?", + "choice_a": "accordion", + "choice_b": "piano", + "choice_c": "guitar", + "choice_d": "bass", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18851 + }, + { + "path": "80782.mp3", + "question": "What instrument is the foundation of this musical number?", + "choice_a": "classicalguitar", + "choice_b": "clarinet", + "choice_c": "sampler", + "choice_d": "piano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18852 + }, + { + "path": "707021.mp3", + "question": "Which instrument is the backbone of this composition?", + "choice_a": "computer", + "choice_b": "drums", + "choice_c": "synthesizer", + "choice_d": "clarinet", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18853 + }, + { + "path": "1064757.mp3", + "question": "Which instrument is the mainstay of this composition?", + "choice_a": "saxophone", + "choice_b": "flute", + "choice_c": "acousticguitar", + "choice_d": "clarinet", + "answer_gt": "saxophone", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18854 + }, + { + "path": "1062504.mp3", + "question": "What's the signature instrument in this piece of music?", + "choice_a": "organ", + "choice_b": "bell", + "choice_c": "synthesizer", + "choice_d": "piano", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18855 + }, + { + "path": "1279772.mp3", + "question": "What instrument is the keystone of this musical piece?", + "choice_a": "saxophone", + "choice_b": "electricguitar", + "choice_c": "voice", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18856 + }, + { + "path": "799923.mp3", + "question": "Which instrument is the core of this composition?", + "choice_a": "classicalguitar", + "choice_b": "harmonica", + "choice_c": "piano", + "choice_d": "voice", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18857 + }, + { + "path": "979833.mp3", + "question": "Which instrument is the focal point of this musical piece?", + "choice_a": "rhodes", + "choice_b": "bell", + "choice_c": "beat", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18858 + }, + { + "path": "1276505.mp3", + "question": "discern the principal instrument in this tune.", + "choice_a": "electricguitar", + "choice_b": "clarinet", + "choice_c": "synthesizer", + "choice_d": "organ", + "answer_gt": "electricguitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18859 + }, + { + "path": "736795.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "cello", + "choice_b": "voice", + "choice_c": "harmonica", + "choice_d": "orchestra", + "answer_gt": "orchestra", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18860 + }, + { + "path": "1316647.mp3", + "question": "What instrument is the anchor of this musical track?", + "choice_a": "electricguitar", + "choice_b": "strings", + "choice_c": "beat", + "choice_d": "violin", + "answer_gt": "beat", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18861 + }, + { + "path": "1348920.mp3", + "question": "Which instrument is most utilized in this piece?", + "choice_a": "acousticguitar", + "choice_b": "classicalguitar", + "choice_c": "pad", + "choice_d": "synthesizer", + "answer_gt": "synthesizer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18862 + }, + { + "path": "1104906.mp3", + "question": "What instrument carries the main theme in this track?", + "choice_a": "orchestra", + "choice_b": "sampler", + "choice_c": "computer", + "choice_d": "harmonica", + "answer_gt": "computer", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18863 + }, + { + "path": "1348662.mp3", + "question": "What instrument is the base of this musical piece?", + "choice_a": "flute", + "choice_b": "pipeorgan", + "choice_c": "piano", + "choice_d": "electricpiano", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18864 + }, + { + "path": "1119925.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "piano", + "choice_b": "drums", + "choice_c": "horn", + "choice_d": "classicalguitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18865 + }, + { + "path": "1244453.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "beat", + "choice_b": "organ", + "choice_c": "voice", + "choice_d": "guitar", + "answer_gt": "guitar", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18866 + }, + { + "path": "1410637.mp3", + "question": "Which instrument is the main contributor to this sound?", + "choice_a": "orchestra", + "choice_b": "cello", + "choice_c": "piano", + "choice_d": "guitar", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18867 + }, + { + "path": "967568.mp3", + "question": "What instrument is the lead in this musical composition?", + "choice_a": "harmonica", + "choice_b": "piano", + "choice_c": "pipeorgan", + "choice_d": "percussion", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18868 + }, + { + "path": "896597.mp3", + "question": "What's the most influential instrument in this piece of music?", + "choice_a": "saxophone", + "choice_b": "piano", + "choice_c": "accordion", + "choice_d": "rhodes", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18869 + }, + { + "path": "1077726.mp3", + "question": "What instrument is the hero of this musical piece?", + "choice_a": "saxophone", + "choice_b": "acousticguitar", + "choice_c": "drummachine", + "choice_d": "harp", + "answer_gt": "harp", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18870 + }, + { + "path": "14023.mp3", + "question": "What's the most pivotal instrument in this music piece?", + "choice_a": "drummachine", + "choice_b": "electricpiano", + "choice_c": "piano", + "choice_d": "saxophone", + "answer_gt": "piano", + "task_name": "Music_Instruments_Classfication", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 18871 + }, + { + "path": "121432.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18872 + }, + { + "path": "116574.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18873 + }, + { + "path": "060601.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18874 + }, + { + "path": "012996.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18875 + }, + { + "path": "014737.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18876 + }, + { + "path": "127181.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18877 + }, + { + "path": "108458.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18878 + }, + { + "path": "014736.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18879 + }, + { + "path": "060995.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18880 + }, + { + "path": "019720.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18881 + }, + { + "path": "068806.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18882 + }, + { + "path": "059925.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18883 + }, + { + "path": "000572.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18884 + }, + { + "path": "053590.mp3", + "question": "What style of music is this?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18885 + }, + { + "path": "110744.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18886 + }, + { + "path": "116195.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18887 + }, + { + "path": "017009.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18888 + }, + { + "path": "071162.mp3", + "question": "what's the genre of this music?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18889 + }, + { + "path": "063457.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18890 + }, + { + "path": "080012.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18891 + }, + { + "path": "098157.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18892 + }, + { + "path": "127051.mp3", + "question": "specify the music genre here?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18893 + }, + { + "path": "015639.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18894 + }, + { + "path": "099093.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18895 + }, + { + "path": "001875.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18896 + }, + { + "path": "127053.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18897 + }, + { + "path": "145454.mp3", + "question": "What's the classification of this music?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18898 + }, + { + "path": "118493.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18899 + }, + { + "path": "116176.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18900 + }, + { + "path": "017008.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18901 + }, + { + "path": "117626.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18902 + }, + { + "path": "093091.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18903 + }, + { + "path": "131458.mp3", + "question": "What style of music is this?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18904 + }, + { + "path": "011508.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Folk", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18905 + }, + { + "path": "082515.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18906 + }, + { + "path": "067779.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18907 + }, + { + "path": "140625.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18908 + }, + { + "path": "115000.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18909 + }, + { + "path": "000565.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18910 + }, + { + "path": "010480.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Pop", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18911 + }, + { + "path": "010443.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18912 + }, + { + "path": "112466.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18913 + }, + { + "path": "112449.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18914 + }, + { + "path": "145750.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18915 + }, + { + "path": "115293.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Pop", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18916 + }, + { + "path": "057732.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18917 + }, + { + "path": "086634.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18918 + }, + { + "path": "005278.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18919 + }, + { + "path": "056526.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18920 + }, + { + "path": "076033.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18921 + }, + { + "path": "010993.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "International", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18922 + }, + { + "path": "075079.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18923 + }, + { + "path": "145513.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "International", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18924 + }, + { + "path": "108461.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18925 + }, + { + "path": "095355.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18926 + }, + { + "path": "012118.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18927 + }, + { + "path": "145747.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18928 + }, + { + "path": "007930.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18929 + }, + { + "path": "074693.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18930 + }, + { + "path": "119826.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18931 + }, + { + "path": "012387.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18932 + }, + { + "path": "053586.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18933 + }, + { + "path": "068752.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18934 + }, + { + "path": "017004.mp3", + "question": "what's the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18935 + }, + { + "path": "004104.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18936 + }, + { + "path": "112474.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18937 + }, + { + "path": "095367.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18938 + }, + { + "path": "000562.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Hip-Hop", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18939 + }, + { + "path": "116578.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18940 + }, + { + "path": "048681.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18941 + }, + { + "path": "112448.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18942 + }, + { + "path": "040238.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18943 + }, + { + "path": "019721.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18944 + }, + { + "path": "082520.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Classical", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18945 + }, + { + "path": "004830.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18946 + }, + { + "path": "001728.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18947 + }, + { + "path": "145887.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18948 + }, + { + "path": "116568.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18949 + }, + { + "path": "117621.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18950 + }, + { + "path": "017011.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18951 + }, + { + "path": "017016.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18952 + }, + { + "path": "060605.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18953 + }, + { + "path": "099079.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18954 + }, + { + "path": "108459.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18955 + }, + { + "path": "076031.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18956 + }, + { + "path": "083786.mp3", + "question": "classify the genre of this music?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18957 + }, + { + "path": "103541.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18958 + }, + { + "path": "005078.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18959 + }, + { + "path": "017015.mp3", + "question": "What's the classification of this music?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18960 + }, + { + "path": "150285.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18961 + }, + { + "path": "015640.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18962 + }, + { + "path": "127186.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18963 + }, + { + "path": "082462.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18964 + }, + { + "path": "096722.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18965 + }, + { + "path": "103537.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18966 + }, + { + "path": "007950.mp3", + "question": "What's the classification of this music?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18967 + }, + { + "path": "112487.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18968 + }, + { + "path": "150284.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18969 + }, + { + "path": "127058.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18970 + }, + { + "path": "119153.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18971 + }, + { + "path": "119154.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18972 + }, + { + "path": "002128.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18973 + }, + { + "path": "131669.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18974 + }, + { + "path": "148610.mp3", + "question": "what's the genre of this music?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18975 + }, + { + "path": "010481.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18976 + }, + { + "path": "011507.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18977 + }, + { + "path": "022558.mp3", + "question": "classify the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18978 + }, + { + "path": "076032.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18979 + }, + { + "path": "127055.mp3", + "question": "what's the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18980 + }, + { + "path": "017713.mp3", + "question": "what's the genre of this music?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18981 + }, + { + "path": "128884.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18982 + }, + { + "path": "005281.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18983 + }, + { + "path": "133102.mp3", + "question": "detect the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18984 + }, + { + "path": "014743.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18985 + }, + { + "path": "110070.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18986 + }, + { + "path": "130752.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18987 + }, + { + "path": "103533.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18988 + }, + { + "path": "080018.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18989 + }, + { + "path": "048680.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18990 + }, + { + "path": "012051.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18991 + }, + { + "path": "140619.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18992 + }, + { + "path": "047505.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18993 + }, + { + "path": "116566.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18994 + }, + { + "path": "010683.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18995 + }, + { + "path": "080016.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18996 + }, + { + "path": "099089.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18997 + }, + { + "path": "148613.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18998 + }, + { + "path": "017440.mp3", + "question": "detect the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 18999 + }, + { + "path": "012991.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Experimental", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19000 + }, + { + "path": "112479.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19001 + }, + { + "path": "059577.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19002 + }, + { + "path": "031819.mp3", + "question": "classify the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19003 + }, + { + "path": "077863.mp3", + "question": "classify the genre of this music?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19004 + }, + { + "path": "095369.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19005 + }, + { + "path": "059907.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19006 + }, + { + "path": "145452.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19007 + }, + { + "path": "119147.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19008 + }, + { + "path": "140611.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19009 + }, + { + "path": "112446.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19010 + }, + { + "path": "068226.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Classical", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19011 + }, + { + "path": "140304.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19012 + }, + { + "path": "146715.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19013 + }, + { + "path": "004108.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19014 + }, + { + "path": "110777.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19015 + }, + { + "path": "014738.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19016 + }, + { + "path": "077867.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19017 + }, + { + "path": "056895.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19018 + }, + { + "path": "062032.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19019 + }, + { + "path": "099091.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19020 + }, + { + "path": "076129.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19021 + }, + { + "path": "110743.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19022 + }, + { + "path": "001048.mp3", + "question": "specify the music genre here?", + "choice_a": "Blues", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19023 + }, + { + "path": "098109.mp3", + "question": "specify the music genre here?", + "choice_a": "Experimental", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19024 + }, + { + "path": "098101.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19025 + }, + { + "path": "001042.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Blues", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19026 + }, + { + "path": "026070.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Jazz", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19027 + }, + { + "path": "076091.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19028 + }, + { + "path": "005277.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19029 + }, + { + "path": "099080.mp3", + "question": "identify the music genre of this track.", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19030 + }, + { + "path": "142361.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19031 + }, + { + "path": "057728.mp3", + "question": "What's the classification of this music?", + "choice_a": "Experimental", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19032 + }, + { + "path": "062030.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Jazz", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19033 + }, + { + "path": "035605.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19034 + }, + { + "path": "040228.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19035 + }, + { + "path": "127331.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19036 + }, + { + "path": "062036.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19037 + }, + { + "path": "116200.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19038 + }, + { + "path": "117625.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19039 + }, + { + "path": "142362.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19040 + }, + { + "path": "075086.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19041 + }, + { + "path": "013814.mp3", + "question": "What style of music is this?", + "choice_a": "International", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19042 + }, + { + "path": "115008.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19043 + }, + { + "path": "068750.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19044 + }, + { + "path": "080020.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19045 + }, + { + "path": "103542.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19046 + }, + { + "path": "001876.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19047 + }, + { + "path": "110076.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19048 + }, + { + "path": "144215.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19049 + }, + { + "path": "116199.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19050 + }, + { + "path": "035603.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19051 + }, + { + "path": "015635.mp3", + "question": "classify the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19052 + }, + { + "path": "140615.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19053 + }, + { + "path": "147021.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19054 + }, + { + "path": "115173.mp3", + "question": "what's the genre of this music?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19055 + }, + { + "path": "014361.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19056 + }, + { + "path": "115294.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "International", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19057 + }, + { + "path": "049477.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19058 + }, + { + "path": "110073.mp3", + "question": "classify the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19059 + }, + { + "path": "035602.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19060 + }, + { + "path": "110084.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19061 + }, + { + "path": "007971.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19062 + }, + { + "path": "063252.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19063 + }, + { + "path": "010381.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19064 + }, + { + "path": "103534.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19065 + }, + { + "path": "082779.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19066 + }, + { + "path": "110754.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19067 + }, + { + "path": "068749.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19068 + }, + { + "path": "106634.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19069 + }, + { + "path": "012486.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19070 + }, + { + "path": "119827.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Country", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19071 + }, + { + "path": "096721.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19072 + }, + { + "path": "110081.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19073 + }, + { + "path": "145516.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19074 + }, + { + "path": "127520.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19075 + }, + { + "path": "110761.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19076 + }, + { + "path": "127519.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19077 + }, + { + "path": "115009.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Experimental", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19078 + }, + { + "path": "010482.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19079 + }, + { + "path": "116177.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19080 + }, + { + "path": "116526.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19081 + }, + { + "path": "004114.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19082 + }, + { + "path": "142359.mp3", + "question": "what's the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19083 + }, + { + "path": "082781.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19084 + }, + { + "path": "121434.mp3", + "question": "classify the genre of this music?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19085 + }, + { + "path": "023010.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19086 + }, + { + "path": "001877.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19087 + }, + { + "path": "012484.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19088 + }, + { + "path": "016125.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19089 + }, + { + "path": "012059.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19090 + }, + { + "path": "128887.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19091 + }, + { + "path": "068342.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19092 + }, + { + "path": "115289.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19093 + }, + { + "path": "110748.mp3", + "question": "What's the classification of this music?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19094 + }, + { + "path": "082786.mp3", + "question": "what's the genre of this music?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19095 + }, + { + "path": "117619.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19096 + }, + { + "path": "071166.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19097 + }, + { + "path": "011614.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19098 + }, + { + "path": "012048.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19099 + }, + { + "path": "014744.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19100 + }, + { + "path": "019718.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19101 + }, + { + "path": "012119.mp3", + "question": "What style of music is this?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19102 + }, + { + "path": "110774.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19103 + }, + { + "path": "117252.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19104 + }, + { + "path": "119830.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19105 + }, + { + "path": "078749.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Blues", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19106 + }, + { + "path": "106631.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19107 + }, + { + "path": "012066.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19108 + }, + { + "path": "012515.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19109 + }, + { + "path": "117624.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19110 + }, + { + "path": "112458.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19111 + }, + { + "path": "035608.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19112 + }, + { + "path": "015647.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19113 + }, + { + "path": "140620.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19114 + }, + { + "path": "110765.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19115 + }, + { + "path": "012390.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19116 + }, + { + "path": "074695.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19117 + }, + { + "path": "116174.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "International", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19118 + }, + { + "path": "030316.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19119 + }, + { + "path": "017442.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19120 + }, + { + "path": "075077.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19121 + }, + { + "path": "110745.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19122 + }, + { + "path": "116410.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19123 + }, + { + "path": "127048.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19124 + }, + { + "path": "017716.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19125 + }, + { + "path": "110768.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19126 + }, + { + "path": "010679.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19127 + }, + { + "path": "082517.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19128 + }, + { + "path": "095360.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19129 + }, + { + "path": "012481.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19130 + }, + { + "path": "103540.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19131 + }, + { + "path": "035609.mp3", + "question": "What style of music is this?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19132 + }, + { + "path": "099085.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19133 + }, + { + "path": "116412.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19134 + }, + { + "path": "018589.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19135 + }, + { + "path": "001864.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19136 + }, + { + "path": "112445.mp3", + "question": "detect the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19137 + }, + { + "path": "059580.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19138 + }, + { + "path": "068812.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19139 + }, + { + "path": "095372.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19140 + }, + { + "path": "147686.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19141 + }, + { + "path": "012068.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19142 + }, + { + "path": "145512.mp3", + "question": "What style of music is this?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19143 + }, + { + "path": "116564.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19144 + }, + { + "path": "041571.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Classical", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19145 + }, + { + "path": "011615.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19146 + }, + { + "path": "096725.mp3", + "question": "What's the classification of this music?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19147 + }, + { + "path": "112478.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19148 + }, + { + "path": "028802.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19149 + }, + { + "path": "018591.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19150 + }, + { + "path": "015646.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19151 + }, + { + "path": "001878.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19152 + }, + { + "path": "099096.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19153 + }, + { + "path": "112440.mp3", + "question": "detect the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19154 + }, + { + "path": "005060.mp3", + "question": "What style of music is this?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19155 + }, + { + "path": "110773.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19156 + }, + { + "path": "059586.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19157 + }, + { + "path": "012047.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Experimental", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19158 + }, + { + "path": "036801.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19159 + }, + { + "path": "056528.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19160 + }, + { + "path": "152199.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19161 + }, + { + "path": "117251.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19162 + }, + { + "path": "099094.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19163 + }, + { + "path": "067778.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19164 + }, + { + "path": "110769.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19165 + }, + { + "path": "110079.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19166 + }, + { + "path": "127057.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19167 + }, + { + "path": "005133.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19168 + }, + { + "path": "066772.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19169 + }, + { + "path": "117622.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19170 + }, + { + "path": "095629.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Classical", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19171 + }, + { + "path": "133492.mp3", + "question": "specify the music genre here?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19172 + }, + { + "path": "082518.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19173 + }, + { + "path": "068809.mp3", + "question": "What's the classification of this music?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19174 + }, + { + "path": "067414.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Rock", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19175 + }, + { + "path": "067774.mp3", + "question": "What style of music is this?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19176 + }, + { + "path": "099090.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19177 + }, + { + "path": "112463.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19178 + }, + { + "path": "051260.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19179 + }, + { + "path": "110753.mp3", + "question": "detect the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19180 + }, + { + "path": "004736.mp3", + "question": "What style of music is this?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19181 + }, + { + "path": "116193.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19182 + }, + { + "path": "017006.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Folk", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19183 + }, + { + "path": "119150.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19184 + }, + { + "path": "116411.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19185 + }, + { + "path": "115174.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19186 + }, + { + "path": "142358.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19187 + }, + { + "path": "057719.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19188 + }, + { + "path": "001868.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19189 + }, + { + "path": "110742.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19190 + }, + { + "path": "015633.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19191 + }, + { + "path": "059689.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19192 + }, + { + "path": "116569.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19193 + }, + { + "path": "082519.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19194 + }, + { + "path": "110074.mp3", + "question": "What's the classification of this music?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19195 + }, + { + "path": "075082.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19196 + }, + { + "path": "001879.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19197 + }, + { + "path": "116198.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19198 + }, + { + "path": "112485.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19199 + }, + { + "path": "072000.mp3", + "question": "What's the classification of this music?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19200 + }, + { + "path": "068344.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19201 + }, + { + "path": "110776.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19202 + }, + { + "path": "018593.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19203 + }, + { + "path": "010686.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19204 + }, + { + "path": "000563.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19205 + }, + { + "path": "115007.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19206 + }, + { + "path": "022560.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19207 + }, + { + "path": "040240.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "International", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19208 + }, + { + "path": "012117.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19209 + }, + { + "path": "112467.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19210 + }, + { + "path": "112452.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19211 + }, + { + "path": "116575.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19212 + }, + { + "path": "072006.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19213 + }, + { + "path": "080017.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19214 + }, + { + "path": "012518.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Pop", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19215 + }, + { + "path": "144212.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19216 + }, + { + "path": "022559.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19217 + }, + { + "path": "110764.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19218 + }, + { + "path": "017020.mp3", + "question": "specify the music genre here?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19219 + }, + { + "path": "115175.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19220 + }, + { + "path": "148002.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19221 + }, + { + "path": "017002.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19222 + }, + { + "path": "098083.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19223 + }, + { + "path": "057734.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19224 + }, + { + "path": "127052.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19225 + }, + { + "path": "122511.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19226 + }, + { + "path": "002126.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19227 + }, + { + "path": "012099.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Blues", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19228 + }, + { + "path": "036998.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19229 + }, + { + "path": "019439.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19230 + }, + { + "path": "117620.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19231 + }, + { + "path": "068823.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19232 + }, + { + "path": "145749.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19233 + }, + { + "path": "133775.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19234 + }, + { + "path": "082467.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19235 + }, + { + "path": "119831.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19236 + }, + { + "path": "145514.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19237 + }, + { + "path": "059585.mp3", + "question": "What's the classification of this music?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19238 + }, + { + "path": "075081.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19239 + }, + { + "path": "106636.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19240 + }, + { + "path": "112476.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19241 + }, + { + "path": "098118.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19242 + }, + { + "path": "084108.mp3", + "question": "classify the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19243 + }, + { + "path": "103545.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19244 + }, + { + "path": "110746.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19245 + }, + { + "path": "036800.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19246 + }, + { + "path": "078748.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Classical", + "choice_b": "Blues", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19247 + }, + { + "path": "110763.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19248 + }, + { + "path": "093090.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19249 + }, + { + "path": "082466.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19250 + }, + { + "path": "108846.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19251 + }, + { + "path": "115290.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19252 + }, + { + "path": "116191.mp3", + "question": "specify the music genre here?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19253 + }, + { + "path": "126240.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19254 + }, + { + "path": "140345.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19255 + }, + { + "path": "031645.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19256 + }, + { + "path": "116203.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19257 + }, + { + "path": "126242.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Pop", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19258 + }, + { + "path": "103530.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19259 + }, + { + "path": "099088.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19260 + }, + { + "path": "060607.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19261 + }, + { + "path": "007956.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19262 + }, + { + "path": "127187.mp3", + "question": "What's the classification of this music?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Jazz", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19263 + }, + { + "path": "116173.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19264 + }, + { + "path": "012513.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19265 + }, + { + "path": "012989.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19266 + }, + { + "path": "110077.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19267 + }, + { + "path": "040230.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19268 + }, + { + "path": "059583.mp3", + "question": "What style of music is this?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19269 + }, + { + "path": "113274.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19270 + }, + { + "path": "004112.mp3", + "question": "detect the genre of this music?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19271 + }, + { + "path": "099076.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19272 + }, + { + "path": "084152.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19273 + }, + { + "path": "115002.mp3", + "question": "classify the genre of this music?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19274 + }, + { + "path": "082464.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19275 + }, + { + "path": "119829.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Instrumental", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "International", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19276 + }, + { + "path": "001045.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Blues", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19277 + }, + { + "path": "110083.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19278 + }, + { + "path": "147700.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Soul-RnB", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Soul-RnB", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19279 + }, + { + "path": "119823.mp3", + "question": "detect the genre of this music?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19280 + }, + { + "path": "145515.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19281 + }, + { + "path": "015642.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19282 + }, + { + "path": "113273.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19283 + }, + { + "path": "073494.mp3", + "question": "detect the genre of this music?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19284 + }, + { + "path": "112469.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19285 + }, + { + "path": "082784.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19286 + }, + { + "path": "131460.mp3", + "question": "specify the music genre here?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19287 + }, + { + "path": "066776.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19288 + }, + { + "path": "017000.mp3", + "question": "specify the music genre here?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19289 + }, + { + "path": "035601.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19290 + }, + { + "path": "068810.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19291 + }, + { + "path": "082516.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Pop", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19292 + }, + { + "path": "001863.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19293 + }, + { + "path": "145777.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19294 + }, + { + "path": "095361.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19295 + }, + { + "path": "016337.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19296 + }, + { + "path": "083785.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19297 + }, + { + "path": "095364.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19298 + }, + { + "path": "010388.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19299 + }, + { + "path": "060828.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19300 + }, + { + "path": "016164.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19301 + }, + { + "path": "082789.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19302 + }, + { + "path": "012052.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19303 + }, + { + "path": "098726.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19304 + }, + { + "path": "053587.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19305 + }, + { + "path": "016127.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19306 + }, + { + "path": "035607.mp3", + "question": "What's the classification of this music?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19307 + }, + { + "path": "145548.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19308 + }, + { + "path": "056527.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19309 + }, + { + "path": "057729.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Folk", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19310 + }, + { + "path": "017721.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19311 + }, + { + "path": "068227.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19312 + }, + { + "path": "011612.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19313 + }, + { + "path": "063456.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19314 + }, + { + "path": "110075.mp3", + "question": "what's the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19315 + }, + { + "path": "103531.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19316 + }, + { + "path": "095370.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19317 + }, + { + "path": "062035.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Country", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19318 + }, + { + "path": "103536.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19319 + }, + { + "path": "080014.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19320 + }, + { + "path": "017013.mp3", + "question": "classify the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19321 + }, + { + "path": "040241.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Folk", + "choice_b": "Pop", + "choice_c": "Country", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19322 + }, + { + "path": "098102.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19323 + }, + { + "path": "150286.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19324 + }, + { + "path": "035610.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19325 + }, + { + "path": "001872.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "International", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19326 + }, + { + "path": "145510.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19327 + }, + { + "path": "110756.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19328 + }, + { + "path": "110770.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19329 + }, + { + "path": "112480.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19330 + }, + { + "path": "112460.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19331 + }, + { + "path": "063251.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19332 + }, + { + "path": "140609.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19333 + }, + { + "path": "059695.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19334 + }, + { + "path": "119146.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19335 + }, + { + "path": "004828.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19336 + }, + { + "path": "017718.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19337 + }, + { + "path": "106630.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19338 + }, + { + "path": "011616.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19339 + }, + { + "path": "110078.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "International", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19340 + }, + { + "path": "084105.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Country", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19341 + }, + { + "path": "093478.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19342 + }, + { + "path": "013328.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19343 + }, + { + "path": "012046.mp3", + "question": "specify the music genre here?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19344 + }, + { + "path": "018588.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19345 + }, + { + "path": "115011.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19346 + }, + { + "path": "098085.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19347 + }, + { + "path": "106633.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19348 + }, + { + "path": "018592.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19349 + }, + { + "path": "010441.mp3", + "question": "What's the classification of this music?", + "choice_a": "Country", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19350 + }, + { + "path": "068822.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19351 + }, + { + "path": "013813.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19352 + }, + { + "path": "108843.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "International", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19353 + }, + { + "path": "128885.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19354 + }, + { + "path": "059693.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19355 + }, + { + "path": "017007.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19356 + }, + { + "path": "001870.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19357 + }, + { + "path": "145455.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19358 + }, + { + "path": "119151.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Country", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19359 + }, + { + "path": "014741.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19360 + }, + { + "path": "046127.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19361 + }, + { + "path": "012179.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19362 + }, + { + "path": "145803.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19363 + }, + { + "path": "011200.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19364 + }, + { + "path": "068807.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19365 + }, + { + "path": "115172.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19366 + }, + { + "path": "010380.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19367 + }, + { + "path": "127059.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19368 + }, + { + "path": "098091.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19369 + }, + { + "path": "000567.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19370 + }, + { + "path": "108842.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19371 + }, + { + "path": "036097.mp3", + "question": "what's the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19372 + }, + { + "path": "017723.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19373 + }, + { + "path": "011503.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19374 + }, + { + "path": "013815.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19375 + }, + { + "path": "099084.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19376 + }, + { + "path": "012514.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19377 + }, + { + "path": "016338.mp3", + "question": "classify the genre of this music?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19378 + }, + { + "path": "010681.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19379 + }, + { + "path": "127182.mp3", + "question": "classify the genre of this music?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19380 + }, + { + "path": "057726.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19381 + }, + { + "path": "014360.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19382 + }, + { + "path": "116579.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19383 + }, + { + "path": "133681.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Country", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19384 + }, + { + "path": "147023.mp3", + "question": "What's the classification of this music?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19385 + }, + { + "path": "040234.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19386 + }, + { + "path": "035615.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19387 + }, + { + "path": "004111.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19388 + }, + { + "path": "112473.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19389 + }, + { + "path": "068234.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19390 + }, + { + "path": "012392.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19391 + }, + { + "path": "099082.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19392 + }, + { + "path": "066777.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19393 + }, + { + "path": "103538.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19394 + }, + { + "path": "126241.mp3", + "question": "What's the classification of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19395 + }, + { + "path": "116570.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19396 + }, + { + "path": "015649.mp3", + "question": "What style of music is this?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19397 + }, + { + "path": "011505.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "International", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19398 + }, + { + "path": "112453.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19399 + }, + { + "path": "082778.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19400 + }, + { + "path": "071200.mp3", + "question": "What style of music is this?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19401 + }, + { + "path": "103543.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19402 + }, + { + "path": "131457.mp3", + "question": "detect the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19403 + }, + { + "path": "078747.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Instrumental", + "choice_b": "Blues", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19404 + }, + { + "path": "040233.mp3", + "question": "specify the music genre here?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19405 + }, + { + "path": "012487.mp3", + "question": "what's the genre of this music?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19406 + }, + { + "path": "076130.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Pop", + "choice_d": "Jazz", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19407 + }, + { + "path": "116202.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19408 + }, + { + "path": "115291.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19409 + }, + { + "path": "012065.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19410 + }, + { + "path": "040244.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19411 + }, + { + "path": "012992.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19412 + }, + { + "path": "012489.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19413 + }, + { + "path": "001043.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Blues", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19414 + }, + { + "path": "014742.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19415 + }, + { + "path": "047774.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19416 + }, + { + "path": "068231.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19417 + }, + { + "path": "010440.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19418 + }, + { + "path": "095368.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19419 + }, + { + "path": "112792.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19420 + }, + { + "path": "145798.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Experimental", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19421 + }, + { + "path": "108456.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19422 + }, + { + "path": "015644.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19423 + }, + { + "path": "059905.mp3", + "question": "detect the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19424 + }, + { + "path": "119825.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19425 + }, + { + "path": "103532.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19426 + }, + { + "path": "011059.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19427 + }, + { + "path": "012067.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19428 + }, + { + "path": "116565.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19429 + }, + { + "path": "040242.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19430 + }, + { + "path": "002120.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19431 + }, + { + "path": "059587.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19432 + }, + { + "path": "110759.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19433 + }, + { + "path": "131456.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19434 + }, + { + "path": "076131.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19435 + }, + { + "path": "035612.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19436 + }, + { + "path": "071203.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19437 + }, + { + "path": "075087.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "International", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19438 + }, + { + "path": "095362.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19439 + }, + { + "path": "063258.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19440 + }, + { + "path": "099077.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19441 + }, + { + "path": "068808.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19442 + }, + { + "path": "004106.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19443 + }, + { + "path": "127060.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19444 + }, + { + "path": "036803.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Experimental", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19445 + }, + { + "path": "112451.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19446 + }, + { + "path": "036799.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19447 + }, + { + "path": "057735.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19448 + }, + { + "path": "005279.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19449 + }, + { + "path": "112456.mp3", + "question": "specify the music genre here?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19450 + }, + { + "path": "062027.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19451 + }, + { + "path": "066774.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19452 + }, + { + "path": "117089.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19453 + }, + { + "path": "059694.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Jazz", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19454 + }, + { + "path": "084736.mp3", + "question": "what's the genre of this music?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19455 + }, + { + "path": "133490.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19456 + }, + { + "path": "112442.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19457 + }, + { + "path": "145511.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19458 + }, + { + "path": "010385.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19459 + }, + { + "path": "012061.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19460 + }, + { + "path": "001046.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Blues", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19461 + }, + { + "path": "012490.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19462 + }, + { + "path": "095365.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19463 + }, + { + "path": "116573.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19464 + }, + { + "path": "119822.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19465 + }, + { + "path": "056522.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19466 + }, + { + "path": "089445.mp3", + "question": "detect the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19467 + }, + { + "path": "112482.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19468 + }, + { + "path": "080011.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19469 + }, + { + "path": "004105.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19470 + }, + { + "path": "005064.mp3", + "question": "What style of music is this?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19471 + }, + { + "path": "012993.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19472 + }, + { + "path": "010111.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19473 + }, + { + "path": "035616.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19474 + }, + { + "path": "012062.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19475 + }, + { + "path": "063253.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19476 + }, + { + "path": "035614.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Instrumental", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19477 + }, + { + "path": "110751.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19478 + }, + { + "path": "059916.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19479 + }, + { + "path": "099081.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19480 + }, + { + "path": "052886.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19481 + }, + { + "path": "121429.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19482 + }, + { + "path": "112481.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19483 + }, + { + "path": "016340.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19484 + }, + { + "path": "122106.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19485 + }, + { + "path": "004109.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19486 + }, + { + "path": "048682.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19487 + }, + { + "path": "112472.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19488 + }, + { + "path": "001874.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19489 + }, + { + "path": "110080.mp3", + "question": "what's the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19490 + }, + { + "path": "077859.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19491 + }, + { + "path": "057721.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19492 + }, + { + "path": "012987.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19493 + }, + { + "path": "112486.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19494 + }, + { + "path": "099092.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19495 + }, + { + "path": "112791.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Jazz", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19496 + }, + { + "path": "116577.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19497 + }, + { + "path": "127185.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Classical", + "choice_b": "Country", + "choice_c": "Pop", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19498 + }, + { + "path": "004113.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19499 + }, + { + "path": "003595.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19500 + }, + { + "path": "014359.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19501 + }, + { + "path": "145477.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19502 + }, + { + "path": "071163.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19503 + }, + { + "path": "117617.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19504 + }, + { + "path": "067780.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19505 + }, + { + "path": "010682.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19506 + }, + { + "path": "012488.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19507 + }, + { + "path": "112793.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19508 + }, + { + "path": "116194.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19509 + }, + { + "path": "117618.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19510 + }, + { + "path": "014733.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19511 + }, + { + "path": "110749.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19512 + }, + { + "path": "082521.mp3", + "question": "What's the classification of this music?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19513 + }, + { + "path": "110747.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19514 + }, + { + "path": "140621.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19515 + }, + { + "path": "086416.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19516 + }, + { + "path": "010992.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19517 + }, + { + "path": "133334.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19518 + }, + { + "path": "012053.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19519 + }, + { + "path": "057724.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19520 + }, + { + "path": "060606.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19521 + }, + { + "path": "017441.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19522 + }, + { + "path": "115006.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19523 + }, + { + "path": "012995.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19524 + }, + { + "path": "110772.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19525 + }, + { + "path": "001881.mp3", + "question": "What style of music is this?", + "choice_a": "Jazz", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19526 + }, + { + "path": "036098.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19527 + }, + { + "path": "075073.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19528 + }, + { + "path": "095628.mp3", + "question": "classify the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19529 + }, + { + "path": "035606.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19530 + }, + { + "path": "000571.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19531 + }, + { + "path": "017720.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19532 + }, + { + "path": "116528.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19533 + }, + { + "path": "075080.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19534 + }, + { + "path": "128888.mp3", + "question": "classify the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19535 + }, + { + "path": "095358.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19536 + }, + { + "path": "112488.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Instrumental", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19537 + }, + { + "path": "145751.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19538 + }, + { + "path": "023063.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19539 + }, + { + "path": "112471.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Experimental", + "choice_b": "Instrumental", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19540 + }, + { + "path": "012050.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19541 + }, + { + "path": "017017.mp3", + "question": "what's the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19542 + }, + { + "path": "099086.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Classical", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19543 + }, + { + "path": "108457.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19544 + }, + { + "path": "144214.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19545 + }, + { + "path": "116201.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "International", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19546 + }, + { + "path": "093089.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "International", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19547 + }, + { + "path": "076030.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Pop", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19548 + }, + { + "path": "010384.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19549 + }, + { + "path": "082785.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19550 + }, + { + "path": "010484.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19551 + }, + { + "path": "002127.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19552 + }, + { + "path": "040237.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19553 + }, + { + "path": "145801.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19554 + }, + { + "path": "127183.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19555 + }, + { + "path": "015641.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19556 + }, + { + "path": "076363.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19557 + }, + { + "path": "145475.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19558 + }, + { + "path": "053588.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19559 + }, + { + "path": "110779.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Pop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19560 + }, + { + "path": "010485.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19561 + }, + { + "path": "122107.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Country", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19562 + }, + { + "path": "010445.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19563 + }, + { + "path": "062021.mp3", + "question": "what's the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19564 + }, + { + "path": "010684.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19565 + }, + { + "path": "005062.mp3", + "question": "specify the music genre here?", + "choice_a": "Country", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19566 + }, + { + "path": "012121.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19567 + }, + { + "path": "112483.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19568 + }, + { + "path": "150288.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Country", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19569 + }, + { + "path": "112477.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19570 + }, + { + "path": "000569.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19571 + }, + { + "path": "040236.mp3", + "question": "specify the music genre here?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19572 + }, + { + "path": "119828.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19573 + }, + { + "path": "013816.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19574 + }, + { + "path": "035613.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19575 + }, + { + "path": "012512.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19576 + }, + { + "path": "003597.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19577 + }, + { + "path": "095371.mp3", + "question": "what's the genre of this music?", + "choice_a": "Classical", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19578 + }, + { + "path": "112790.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19579 + }, + { + "path": "095356.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Country", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19580 + }, + { + "path": "119152.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19581 + }, + { + "path": "112444.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19582 + }, + { + "path": "116571.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19583 + }, + { + "path": "012374.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19584 + }, + { + "path": "112459.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19585 + }, + { + "path": "099095.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19586 + }, + { + "path": "075075.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19587 + }, + { + "path": "012485.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19588 + }, + { + "path": "010108.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19589 + }, + { + "path": "051779.mp3", + "question": "detect the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19590 + }, + { + "path": "017001.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19591 + }, + { + "path": "068235.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19592 + }, + { + "path": "057736.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19593 + }, + { + "path": "068340.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Classical", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19594 + }, + { + "path": "103535.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Folk", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19595 + }, + { + "path": "005275.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19596 + }, + { + "path": "127049.mp3", + "question": "detect the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19597 + }, + { + "path": "012058.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19598 + }, + { + "path": "110757.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19599 + }, + { + "path": "113279.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19600 + }, + { + "path": "035394.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19601 + }, + { + "path": "115005.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19602 + }, + { + "path": "062020.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19603 + }, + { + "path": "060603.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19604 + }, + { + "path": "083787.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Hip-Hop", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19605 + }, + { + "path": "112457.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19606 + }, + { + "path": "112441.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19607 + }, + { + "path": "116572.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19608 + }, + { + "path": "116567.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "International", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19609 + }, + { + "path": "011613.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19610 + }, + { + "path": "148608.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19611 + }, + { + "path": "010386.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19612 + }, + { + "path": "147020.mp3", + "question": "What's the classification of this music?", + "choice_a": "Folk", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19613 + }, + { + "path": "010483.mp3", + "question": "detect the genre of this music?", + "choice_a": "Country", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19614 + }, + { + "path": "004110.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19615 + }, + { + "path": "107802.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Classical", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19616 + }, + { + "path": "116175.mp3", + "question": "What's the classification of this music?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19617 + }, + { + "path": "036804.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19618 + }, + { + "path": "140624.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19619 + }, + { + "path": "062023.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19620 + }, + { + "path": "035611.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19621 + }, + { + "path": "115001.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19622 + }, + { + "path": "116192.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Pop", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19623 + }, + { + "path": "115170.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Rock", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19624 + }, + { + "path": "067775.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19625 + }, + { + "path": "145517.mp3", + "question": "detect the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Classical", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19626 + }, + { + "path": "112484.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Folk", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19627 + }, + { + "path": "015645.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19628 + }, + { + "path": "145800.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19629 + }, + { + "path": "012482.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Hip-Hop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19630 + }, + { + "path": "046720.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19631 + }, + { + "path": "010685.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19632 + }, + { + "path": "140623.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19633 + }, + { + "path": "118505.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19634 + }, + { + "path": "110767.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19635 + }, + { + "path": "040245.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Country", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19636 + }, + { + "path": "010442.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19637 + }, + { + "path": "127184.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19638 + }, + { + "path": "117616.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19639 + }, + { + "path": "076128.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19640 + }, + { + "path": "020056.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19641 + }, + { + "path": "017717.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19642 + }, + { + "path": "095363.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19643 + }, + { + "path": "082787.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19644 + }, + { + "path": "010447.mp3", + "question": "What's the classification of this music?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19645 + }, + { + "path": "145457.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Instrumental", + "choice_b": "Pop", + "choice_c": "Jazz", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19646 + }, + { + "path": "059906.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19647 + }, + { + "path": "010382.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19648 + }, + { + "path": "086415.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19649 + }, + { + "path": "068341.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19650 + }, + { + "path": "004829.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19651 + }, + { + "path": "056523.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19652 + }, + { + "path": "131459.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19653 + }, + { + "path": "080019.mp3", + "question": "What style of music is this?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19654 + }, + { + "path": "015637.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Instrumental", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19655 + }, + { + "path": "019719.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19656 + }, + { + "path": "147022.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "International", + "choice_b": "Classical", + "choice_c": "Experimental", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19657 + }, + { + "path": "002129.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Country", + "choice_b": "Folk", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19658 + }, + { + "path": "010383.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19659 + }, + { + "path": "040231.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19660 + }, + { + "path": "060476.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19661 + }, + { + "path": "057725.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Experimental", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19662 + }, + { + "path": "005061.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19663 + }, + { + "path": "010444.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19664 + }, + { + "path": "115287.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19665 + }, + { + "path": "151995.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19666 + }, + { + "path": "010130.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19667 + }, + { + "path": "068229.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19668 + }, + { + "path": "019724.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19669 + }, + { + "path": "133491.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19670 + }, + { + "path": "046718.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19671 + }, + { + "path": "066775.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Folk", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Electronic", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19672 + }, + { + "path": "099083.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19673 + }, + { + "path": "083053.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19674 + }, + { + "path": "110778.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19675 + }, + { + "path": "062037.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19676 + }, + { + "path": "062033.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19677 + }, + { + "path": "011504.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19678 + }, + { + "path": "035604.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19679 + }, + { + "path": "045144.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19680 + }, + { + "path": "098723.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Instrumental", + "choice_b": "Hip-Hop", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19681 + }, + { + "path": "122948.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19682 + }, + { + "path": "110072.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19683 + }, + { + "path": "003842.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Pop", + "choice_b": "Blues", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19684 + }, + { + "path": "056524.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19685 + }, + { + "path": "075074.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Hip-Hop", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19686 + }, + { + "path": "040232.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Pop", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19687 + }, + { + "path": "080013.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19688 + }, + { + "path": "013812.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "International", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19689 + }, + { + "path": "067777.mp3", + "question": "What's the classification of this music?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19690 + }, + { + "path": "001044.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Blues", + "answer_gt": "Blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19691 + }, + { + "path": "142360.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19692 + }, + { + "path": "012063.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Experimental", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19693 + }, + { + "path": "057723.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19694 + }, + { + "path": "121435.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Jazz", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19695 + }, + { + "path": "012120.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19696 + }, + { + "path": "127061.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Instrumental", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19697 + }, + { + "path": "140616.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Pop", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19698 + }, + { + "path": "082465.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19699 + }, + { + "path": "148612.mp3", + "question": "what's the genre of this music?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19700 + }, + { + "path": "003596.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19701 + }, + { + "path": "017715.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19702 + }, + { + "path": "053589.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Folk", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19703 + }, + { + "path": "012045.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19704 + }, + { + "path": "059584.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19705 + }, + { + "path": "072002.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19706 + }, + { + "path": "145476.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Country", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19707 + }, + { + "path": "010439.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Jazz", + "choice_b": "Country", + "choice_c": "Folk", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19708 + }, + { + "path": "062034.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19709 + }, + { + "path": "131455.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19710 + }, + { + "path": "005065.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Rock", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Classical", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19711 + }, + { + "path": "062025.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19712 + }, + { + "path": "004707.mp3", + "question": "what's the genre of this music?", + "choice_a": "Pop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19713 + }, + { + "path": "075089.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19714 + }, + { + "path": "059578.mp3", + "question": "identify the music genre of this track.", + "choice_a": "Hip-Hop", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19715 + }, + { + "path": "017014.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19716 + }, + { + "path": "071953.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19717 + }, + { + "path": "016339.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19718 + }, + { + "path": "112475.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19719 + }, + { + "path": "047504.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19720 + }, + { + "path": "117253.mp3", + "question": "detect the genre of this music?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Pop", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19721 + }, + { + "path": "063250.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19722 + }, + { + "path": "098078.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19723 + }, + { + "path": "128441.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19724 + }, + { + "path": "068233.mp3", + "question": "detect the genre of this music?", + "choice_a": "Classical", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19725 + }, + { + "path": "106637.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19726 + }, + { + "path": "108847.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19727 + }, + { + "path": "106638.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19728 + }, + { + "path": "083784.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19729 + }, + { + "path": "015634.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19730 + }, + { + "path": "082782.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19731 + }, + { + "path": "011506.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19732 + }, + { + "path": "106632.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Classical", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19733 + }, + { + "path": "095630.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19734 + }, + { + "path": "103539.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Classical", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19735 + }, + { + "path": "012997.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19736 + }, + { + "path": "062026.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19737 + }, + { + "path": "072003.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Country", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19738 + }, + { + "path": "017712.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19739 + }, + { + "path": "110071.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Instrumental", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19740 + }, + { + "path": "082469.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19741 + }, + { + "path": "121433.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19742 + }, + { + "path": "140618.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19743 + }, + { + "path": "131461.mp3", + "question": "What's the classification of this music?", + "choice_a": "Jazz", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19744 + }, + { + "path": "144213.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19745 + }, + { + "path": "075084.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "International", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19746 + }, + { + "path": "131462.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Jazz", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19747 + }, + { + "path": "026071.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "International", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19748 + }, + { + "path": "072005.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Electronic", + "choice_b": "Pop", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19749 + }, + { + "path": "068230.mp3", + "question": "detect the genre of this music?", + "choice_a": "Jazz", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Electronic", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19750 + }, + { + "path": "148611.mp3", + "question": "classify the genre of this music?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19751 + }, + { + "path": "110771.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19752 + }, + { + "path": "010678.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Experimental", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19753 + }, + { + "path": "017005.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Jazz", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19754 + }, + { + "path": "012049.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19755 + }, + { + "path": "074689.mp3", + "question": "specify the music genre here?", + "choice_a": "Pop", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Rock", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19756 + }, + { + "path": "012511.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19757 + }, + { + "path": "072007.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19758 + }, + { + "path": "040239.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19759 + }, + { + "path": "145456.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Experimental", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19760 + }, + { + "path": "001727.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Jazz", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19761 + }, + { + "path": "075090.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "Classical", + "choice_b": "International", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19762 + }, + { + "path": "116409.mp3", + "question": "specify the music genre here?", + "choice_a": "Classical", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19763 + }, + { + "path": "074694.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "Pop", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19764 + }, + { + "path": "110082.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Pop", + "choice_b": "Jazz", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19765 + }, + { + "path": "122536.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Pop", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19766 + }, + { + "path": "110086.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Jazz", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19767 + }, + { + "path": "014734.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Country", + "choice_d": "Experimental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19768 + }, + { + "path": "004708.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Jazz", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19769 + }, + { + "path": "132117.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19770 + }, + { + "path": "052884.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19771 + }, + { + "path": "035398.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19772 + }, + { + "path": "018042.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Pop", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19773 + }, + { + "path": "125685.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Instrumental", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19774 + }, + { + "path": "110760.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Jazz", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19775 + }, + { + "path": "017019.mp3", + "question": "specify the music genre here?", + "choice_a": "Country", + "choice_b": "Folk", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19776 + }, + { + "path": "112443.mp3", + "question": "classify the genre of this music?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19777 + }, + { + "path": "015638.mp3", + "question": "what's the genre of this music?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19778 + }, + { + "path": "140612.mp3", + "question": "specify the music genre here?", + "choice_a": "Hip-Hop", + "choice_b": "Classical", + "choice_c": "Jazz", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19779 + }, + { + "path": "005280.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19780 + }, + { + "path": "098111.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Electronic", + "choice_b": "International", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19781 + }, + { + "path": "057720.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Pop", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19782 + }, + { + "path": "010387.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19783 + }, + { + "path": "036099.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19784 + }, + { + "path": "014745.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19785 + }, + { + "path": "040229.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Rock", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19786 + }, + { + "path": "116197.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19787 + }, + { + "path": "000568.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Rock", + "choice_b": "International", + "choice_c": "Classical", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19788 + }, + { + "path": "062022.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Rock", + "choice_b": "Instrumental", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19789 + }, + { + "path": "057731.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19790 + }, + { + "path": "152056.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19791 + }, + { + "path": "029266.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Jazz", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19792 + }, + { + "path": "010389.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19793 + }, + { + "path": "059579.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19794 + }, + { + "path": "071818.mp3", + "question": "What style of music is this?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19795 + }, + { + "path": "099074.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "Pop", + "choice_b": "Experimental", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19796 + }, + { + "path": "093479.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Experimental", + "choice_b": "Classical", + "choice_c": "Rock", + "choice_d": "Folk", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19797 + }, + { + "path": "095357.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19798 + }, + { + "path": "112464.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "Pop", + "choice_b": "International", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19799 + }, + { + "path": "074690.mp3", + "question": "detect the genre of this music?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Rock", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19800 + }, + { + "path": "112447.mp3", + "question": "What's the classification of this music?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Classical", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19801 + }, + { + "path": "015643.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Instrumental", + "choice_c": "Electronic", + "choice_d": "Country", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19802 + }, + { + "path": "075083.mp3", + "question": "What style of music is this?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19803 + }, + { + "path": "018040.mp3", + "question": "detect the genre of this music?", + "choice_a": "Instrumental", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19804 + }, + { + "path": "084107.mp3", + "question": "detect the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19805 + }, + { + "path": "119149.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Electronic", + "choice_b": "Experimental", + "choice_c": "Instrumental", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19806 + }, + { + "path": "121431.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Jazz", + "choice_d": "Experimental", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19807 + }, + { + "path": "127054.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "Rock", + "choice_b": "Country", + "choice_c": "Experimental", + "choice_d": "Folk", + "answer_gt": "Country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19808 + }, + { + "path": "005276.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19809 + }, + { + "path": "127189.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "International", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Classical", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19810 + }, + { + "path": "108460.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "International", + "choice_b": "Pop", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19811 + }, + { + "path": "140626.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Rock", + "choice_b": "Pop", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19812 + }, + { + "path": "112455.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19813 + }, + { + "path": "059582.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Rock", + "choice_b": "Classical", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19814 + }, + { + "path": "099075.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19815 + }, + { + "path": "068228.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Hip-Hop", + "choice_b": "Country", + "choice_c": "Classical", + "choice_d": "Jazz", + "answer_gt": "Jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19816 + }, + { + "path": "071950.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Electronic", + "choice_b": "Classical", + "choice_c": "Folk", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19817 + }, + { + "path": "075085.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Jazz", + "choice_b": "Instrumental", + "choice_c": "International", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19818 + }, + { + "path": "110085.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19819 + }, + { + "path": "112450.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19820 + }, + { + "path": "071999.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "Electronic", + "choice_b": "Instrumental", + "choice_c": "Country", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19821 + }, + { + "path": "110762.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Electronic", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19822 + }, + { + "path": "016161.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Folk", + "choice_b": "Experimental", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19823 + }, + { + "path": "145458.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19824 + }, + { + "path": "074692.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Pop", + "choice_b": "Classical", + "choice_c": "International", + "choice_d": "Instrumental", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19825 + }, + { + "path": "062031.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Classical", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19826 + }, + { + "path": "057727.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Country", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19827 + }, + { + "path": "098080.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19828 + }, + { + "path": "119145.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "Classical", + "choice_b": "Folk", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19829 + }, + { + "path": "076029.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Country", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19830 + }, + { + "path": "122510.mp3", + "question": "specify the music genre here?", + "choice_a": "Folk", + "choice_b": "International", + "choice_c": "Electronic", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19831 + }, + { + "path": "115288.mp3", + "question": "detect the genre of this music?", + "choice_a": "Folk", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Jazz", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19832 + }, + { + "path": "112468.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Experimental", + "choice_b": "Electronic", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19833 + }, + { + "path": "110758.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "International", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Instrumental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19834 + }, + { + "path": "060604.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19835 + }, + { + "path": "017714.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Jazz", + "choice_c": "Rock", + "choice_d": "Hip-Hop", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19836 + }, + { + "path": "013673.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Rock", + "choice_b": "Electronic", + "choice_c": "Pop", + "choice_d": "Folk", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19837 + }, + { + "path": "040243.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "International", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19838 + }, + { + "path": "012990.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19839 + }, + { + "path": "112465.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "Experimental", + "choice_b": "Country", + "choice_c": "Hip-Hop", + "choice_d": "Jazz", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19840 + }, + { + "path": "112454.mp3", + "question": "specify the music genre here?", + "choice_a": "Experimental", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19841 + }, + { + "path": "063257.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "Classical", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Folk", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19842 + }, + { + "path": "145805.mp3", + "question": "specify the music genre here?", + "choice_a": "Pop", + "choice_b": "Electronic", + "choice_c": "Hip-Hop", + "choice_d": "International", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19843 + }, + { + "path": "098099.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "Electronic", + "choice_b": "Folk", + "choice_c": "Classical", + "choice_d": "Rock", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19844 + }, + { + "path": "075076.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "Rock", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19845 + }, + { + "path": "110775.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "Hip-Hop", + "choice_b": "Rock", + "choice_c": "Country", + "choice_d": "Folk", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19846 + }, + { + "path": "113262.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Experimental", + "choice_c": "International", + "choice_d": "Pop", + "answer_gt": "International", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19847 + }, + { + "path": "148606.mp3", + "question": "What's the classification of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19848 + }, + { + "path": "062029.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Hip-Hop", + "choice_b": "Folk", + "choice_c": "Experimental", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19849 + }, + { + "path": "075088.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Instrumental", + "choice_b": "Jazz", + "choice_c": "Pop", + "choice_d": "Hip-Hop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19850 + }, + { + "path": "084104.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "Hip-Hop", + "choice_b": "Pop", + "choice_c": "Classical", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19851 + }, + { + "path": "014735.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "Folk", + "choice_b": "Hip-Hop", + "choice_c": "Experimental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19852 + }, + { + "path": "112462.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Experimental", + "choice_b": "Hip-Hop", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19853 + }, + { + "path": "110755.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Rock", + "choice_d": "Country", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19854 + }, + { + "path": "098128.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Classical", + "choice_b": "Country", + "choice_c": "Electronic", + "choice_d": "Hip-Hop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19855 + }, + { + "path": "148609.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "Country", + "choice_b": "Pop", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19856 + }, + { + "path": "030000.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "Hip-Hop", + "choice_b": "Pop", + "choice_c": "Electronic", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19857 + }, + { + "path": "140610.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Experimental", + "choice_d": "International", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19858 + }, + { + "path": "106639.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "Country", + "choice_b": "Electronic", + "choice_c": "Rock", + "choice_d": "Experimental", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19859 + }, + { + "path": "151988.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "International", + "choice_b": "Experimental", + "choice_c": "Folk", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19860 + }, + { + "path": "095366.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "Folk", + "choice_b": "Electronic", + "choice_c": "Classical", + "choice_d": "Pop", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19861 + }, + { + "path": "115004.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Instrumental", + "choice_d": "Folk", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19862 + }, + { + "path": "115003.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Electronic", + "choice_b": "Hip-Hop", + "choice_c": "Country", + "choice_d": "Instrumental", + "answer_gt": "Instrumental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19863 + }, + { + "path": "098112.mp3", + "question": "what's the genre of this music?", + "choice_a": "Electronic", + "choice_b": "Rock", + "choice_c": "Folk", + "choice_d": "Classical", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19864 + }, + { + "path": "041810.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "Pop", + "choice_b": "Country", + "choice_c": "Instrumental", + "choice_d": "Rock", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19865 + }, + { + "path": "000566.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "Rock", + "choice_b": "Jazz", + "choice_c": "Folk", + "choice_d": "Instrumental", + "answer_gt": "Rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19866 + }, + { + "path": "057733.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "Jazz", + "choice_b": "Experimental", + "choice_c": "Pop", + "choice_d": "Rock", + "answer_gt": "Experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19867 + }, + { + "path": "075078.mp3", + "question": "specify the music genre here?", + "choice_a": "Hip-Hop", + "choice_b": "Jazz", + "choice_c": "Electronic", + "choice_d": "Pop", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19868 + }, + { + "path": "140613.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "Instrumental", + "choice_b": "Pop", + "choice_c": "Hip-Hop", + "choice_d": "Experimental", + "answer_gt": "Hip-Hop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19869 + }, + { + "path": "012060.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "Jazz", + "choice_b": "Folk", + "choice_c": "Hip-Hop", + "choice_d": "Country", + "answer_gt": "Folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19870 + }, + { + "path": "022561.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "Country", + "choice_b": "Instrumental", + "choice_c": "Rock", + "choice_d": "Electronic", + "answer_gt": "Electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "fma", + "uniq_id": 19871 + }, + { + "path": "1167078.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "classical", + "choice_b": "hiphop", + "choice_c": "indie", + "choice_d": "dance", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19872 + }, + { + "path": "1214582.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "rocknroll", + "choice_b": "soul", + "choice_c": "instrumentalpop", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19873 + }, + { + "path": "1167077.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "classical", + "choice_b": "atmospheric", + "choice_c": "electropop", + "choice_d": "latin", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19874 + }, + { + "path": "111957.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "reggae", + "choice_b": "rap", + "choice_c": "easylistening", + "choice_d": "minimal", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19875 + }, + { + "path": "1093497.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "rap", + "choice_b": "hiphop", + "choice_c": "country", + "choice_d": "jazz", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19876 + }, + { + "path": "979860.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "ambient", + "choice_b": "easylistening", + "choice_c": "breakbeat", + "choice_d": "industrial", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19877 + }, + { + "path": "295917.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "idm", + "choice_b": "industrial", + "choice_c": "popfolk", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19878 + }, + { + "path": "1251474.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "house", + "choice_b": "alternative", + "choice_c": "reggae", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19879 + }, + { + "path": "1026818.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "soundtrack", + "choice_b": "world", + "choice_c": "techno", + "choice_d": "trance", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19880 + }, + { + "path": "764.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "swing", + "choice_b": "pop", + "choice_c": "country", + "choice_d": "techno", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19881 + }, + { + "path": "283380.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "latin", + "choice_b": "house", + "choice_c": "electronic", + "choice_d": "country", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19882 + }, + { + "path": "1051716.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "industrial", + "choice_b": "folk", + "choice_c": "hard", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19883 + }, + { + "path": "588709.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "chillout", + "choice_b": "punkrock", + "choice_c": "acidjazz", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19884 + }, + { + "path": "965319.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "swing", + "choice_b": "classical", + "choice_c": "soul", + "choice_d": "reggae", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19885 + }, + { + "path": "208299.mp3", + "question": "classify the genre of this music?", + "choice_a": "rock", + "choice_b": "trance", + "choice_c": "industrial", + "choice_d": "folk", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19886 + }, + { + "path": "965303.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "classical", + "choice_b": "newwave", + "choice_c": "darkambient", + "choice_d": "ethno", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19887 + }, + { + "path": "901995.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "soundtrack", + "choice_b": "newwave", + "choice_c": "electronic", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19888 + }, + { + "path": "141086.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "experimental", + "choice_b": "newwave", + "choice_c": "progressive", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19889 + }, + { + "path": "335254.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "chillout", + "choice_b": "chanson", + "choice_c": "classical", + "choice_d": "rap", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19890 + }, + { + "path": "42316.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "darkambient", + "choice_b": "experimental", + "choice_c": "rock", + "choice_d": "house", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19891 + }, + { + "path": "989713.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "metal", + "choice_b": "funk", + "choice_c": "trance", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19892 + }, + { + "path": "979864.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "electropop", + "choice_b": "metal", + "choice_c": "drumnbass", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19893 + }, + { + "path": "245760.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "metal", + "choice_b": "dance", + "choice_c": "electronic", + "choice_d": "idm", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19894 + }, + { + "path": "1160029.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "triphop", + "choice_b": "swing", + "choice_c": "psychedelic", + "choice_d": "folk", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19895 + }, + { + "path": "147198.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "reggae", + "choice_b": "punkrock", + "choice_c": "world", + "choice_d": "rock", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19896 + }, + { + "path": "14022.mp3", + "question": "What's the classification of this music?", + "choice_a": "classical", + "choice_b": "latin", + "choice_c": "atmospheric", + "choice_d": "drumnbass", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19897 + }, + { + "path": "1393141.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "chillout", + "choice_b": "breakbeat", + "choice_c": "techno", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19898 + }, + { + "path": "1051711.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "acidjazz", + "choice_b": "rock", + "choice_c": "easylistening", + "choice_d": "instrumentalpop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19899 + }, + { + "path": "1105855.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "bossanova", + "choice_b": "classical", + "choice_c": "rocknroll", + "choice_d": "newage", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19900 + }, + { + "path": "1302998.mp3", + "question": "What's the classification of this music?", + "choice_a": "ethno", + "choice_b": "blues", + "choice_c": "easylistening", + "choice_d": "soundtrack", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19901 + }, + { + "path": "121147.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "newwave", + "choice_b": "world", + "choice_c": "atmospheric", + "choice_d": "hiphop", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19902 + }, + { + "path": "42344.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "dance", + "choice_b": "rock", + "choice_c": "folk", + "choice_d": "indie", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19903 + }, + { + "path": "111956.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "ethno", + "choice_b": "pop", + "choice_c": "trance", + "choice_d": "soul", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19904 + }, + { + "path": "288365.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "metal", + "choice_b": "electronica", + "choice_c": "classical", + "choice_d": "industrial", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19905 + }, + { + "path": "1268.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "metal", + "choice_b": "classical", + "choice_c": "chillout", + "choice_d": "popfolk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19906 + }, + { + "path": "842791.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "classical", + "choice_b": "drumnbass", + "choice_c": "electronica", + "choice_d": "triphop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19907 + }, + { + "path": "1094517.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "ethno", + "choice_b": "techno", + "choice_c": "trance", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19908 + }, + { + "path": "179545.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "instrumentalpop", + "choice_b": "rocknroll", + "choice_c": "electronic", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19909 + }, + { + "path": "1044827.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "psychedelic", + "choice_b": "metal", + "choice_c": "orchestral", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19910 + }, + { + "path": "639276.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "trance", + "choice_b": "electronic", + "choice_c": "industrial", + "choice_d": "minimal", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19911 + }, + { + "path": "1293518.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "triphop", + "choice_b": "newwave", + "choice_c": "pop", + "choice_d": "acidjazz", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19912 + }, + { + "path": "368180.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "electronica", + "choice_b": "poprock", + "choice_c": "progressive", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19913 + }, + { + "path": "182013.mp3", + "question": "detect the genre of this music?", + "choice_a": "triphop", + "choice_b": "chillout", + "choice_c": "electronic", + "choice_d": "hiphop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19914 + }, + { + "path": "973732.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "drumnbass", + "choice_b": "ethno", + "choice_c": "hard", + "choice_d": "funk", + "answer_gt": "funk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19915 + }, + { + "path": "1350546.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "hard", + "choice_b": "rock", + "choice_c": "rap", + "choice_d": "electronic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19916 + }, + { + "path": "914771.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "folk", + "choice_b": "psychedelic", + "choice_c": "progressive", + "choice_d": "soul", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19917 + }, + { + "path": "123742.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "experimental", + "choice_b": "house", + "choice_c": "breakbeat", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19918 + }, + { + "path": "1232307.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "ethno", + "choice_b": "rap", + "choice_c": "breakbeat", + "choice_d": "experimental", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19919 + }, + { + "path": "524235.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "bossanova", + "choice_b": "acidjazz", + "choice_c": "electronic", + "choice_d": "newwave", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19920 + }, + { + "path": "1329121.mp3", + "question": "What style of music is this?", + "choice_a": "instrumentalpop", + "choice_b": "minimal", + "choice_c": "triphop", + "choice_d": "funk", + "answer_gt": "funk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19921 + }, + { + "path": "1085249.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "classical", + "choice_b": "progressive", + "choice_c": "psychedelic", + "choice_d": "electronic", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19922 + }, + { + "path": "1348113.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "darkambient", + "choice_b": "classical", + "choice_c": "alternative", + "choice_d": "electronic", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19923 + }, + { + "path": "208287.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "easylistening", + "choice_b": "jazz", + "choice_c": "electronic", + "choice_d": "metal", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19924 + }, + { + "path": "1147241.mp3", + "question": "What's the classification of this music?", + "choice_a": "poprock", + "choice_b": "country", + "choice_c": "folk", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19925 + }, + { + "path": "1159856.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "trance", + "choice_b": "swing", + "choice_c": "instrumentalpop", + "choice_d": "indie", + "answer_gt": "indie", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19926 + }, + { + "path": "362494.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "pop", + "choice_b": "funk", + "choice_c": "soundtrack", + "choice_d": "progressive", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19927 + }, + { + "path": "1321124.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "world", + "choice_b": "classical", + "choice_c": "acidjazz", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19928 + }, + { + "path": "431436.mp3", + "question": "what's the genre of this music?", + "choice_a": "indie", + "choice_b": "triphop", + "choice_c": "minimal", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19929 + }, + { + "path": "133673.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "celtic", + "choice_b": "electronic", + "choice_c": "easylistening", + "choice_d": "rocknroll", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19930 + }, + { + "path": "1395313.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "hard", + "choice_b": "swing", + "choice_c": "country", + "choice_d": "latin", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19931 + }, + { + "path": "1145058.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "alternative", + "choice_b": "instrumentalpop", + "choice_c": "hard", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19932 + }, + { + "path": "165328.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "electronic", + "choice_b": "bossanova", + "choice_c": "drumnbass", + "choice_d": "experimental", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19933 + }, + { + "path": "393252.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "classical", + "choice_b": "world", + "choice_c": "ethno", + "choice_d": "easylistening", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19934 + }, + { + "path": "1286486.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "trance", + "choice_b": "classical", + "choice_c": "newage", + "choice_d": "minimal", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19935 + }, + { + "path": "1352577.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "soul", + "choice_b": "soundtrack", + "choice_c": "reggae", + "choice_d": "acidjazz", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19936 + }, + { + "path": "1420710.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "jazz", + "choice_b": "latin", + "choice_c": "psychedelic", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19937 + }, + { + "path": "1348124.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "soundtrack", + "choice_b": "ambient", + "choice_c": "bossanova", + "choice_d": "orchestral", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19938 + }, + { + "path": "1350975.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "soundtrack", + "choice_b": "country", + "choice_c": "techno", + "choice_d": "latin", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19939 + }, + { + "path": "224794.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "progressive", + "choice_b": "electronic", + "choice_c": "latin", + "choice_d": "house", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19940 + }, + { + "path": "31505.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "pop", + "choice_b": "soundtrack", + "choice_c": "jazz", + "choice_d": "experimental", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19941 + }, + { + "path": "1263.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "newage", + "choice_b": "darkambient", + "choice_c": "funk", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19942 + }, + { + "path": "519178.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "alternative", + "choice_b": "hiphop", + "choice_c": "atmospheric", + "choice_d": "techno", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19943 + }, + { + "path": "393240.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "country", + "choice_b": "folk", + "choice_c": "classical", + "choice_d": "progressive", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19944 + }, + { + "path": "1180362.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electropop", + "choice_b": "psychedelic", + "choice_c": "poprock", + "choice_d": "world", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19945 + }, + { + "path": "1094574.mp3", + "question": "What's the classification of this music?", + "choice_a": "bossanova", + "choice_b": "alternative", + "choice_c": "folk", + "choice_d": "hardrock", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19946 + }, + { + "path": "979839.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "dance", + "choice_b": "ambient", + "choice_c": "rnb", + "choice_d": "funk", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19947 + }, + { + "path": "287980.mp3", + "question": "What style of music is this?", + "choice_a": "house", + "choice_b": "minimal", + "choice_c": "rocknroll", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19948 + }, + { + "path": "25262.mp3", + "question": "what's the genre of this music?", + "choice_a": "electronica", + "choice_b": "soundtrack", + "choice_c": "house", + "choice_d": "psychedelic", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19949 + }, + { + "path": "347228.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "country", + "choice_b": "idm", + "choice_c": "rock", + "choice_d": "drumnbass", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19950 + }, + { + "path": "1419774.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "electronica", + "choice_b": "techno", + "choice_c": "ambient", + "choice_d": "ethno", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19951 + }, + { + "path": "1096098.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "trance", + "choice_b": "rock", + "choice_c": "classical", + "choice_d": "newwave", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19952 + }, + { + "path": "1133071.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "newwave", + "choice_b": "instrumentalpop", + "choice_c": "electronica", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19953 + }, + { + "path": "1076566.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "hiphop", + "choice_b": "world", + "choice_c": "folk", + "choice_d": "metal", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19954 + }, + { + "path": "201979.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "minimal", + "choice_b": "latin", + "choice_c": "electronic", + "choice_d": "psychedelic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19955 + }, + { + "path": "973729.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "funk", + "choice_b": "hard", + "choice_c": "latin", + "choice_d": "electronic", + "answer_gt": "funk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19956 + }, + { + "path": "1350830.mp3", + "question": "detect the genre of this music?", + "choice_a": "chillout", + "choice_b": "alternative", + "choice_c": "soundtrack", + "choice_d": "acidjazz", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19957 + }, + { + "path": "1109352.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "experimental", + "choice_b": "ethno", + "choice_c": "electronic", + "choice_d": "hard", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19958 + }, + { + "path": "1031108.mp3", + "question": "classify the genre of this music?", + "choice_a": "industrial", + "choice_b": "blues", + "choice_c": "electronic", + "choice_d": "world", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19959 + }, + { + "path": "969340.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "rnb", + "choice_b": "indie", + "choice_c": "orchestral", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19960 + }, + { + "path": "1348864.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "poprock", + "choice_b": "electropop", + "choice_c": "country", + "choice_d": "breakbeat", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19961 + }, + { + "path": "1393663.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "hardrock", + "choice_b": "house", + "choice_c": "bossanova", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19962 + }, + { + "path": "856982.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "rock", + "choice_b": "electronic", + "choice_c": "hiphop", + "choice_d": "ethno", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19963 + }, + { + "path": "1329119.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "country", + "choice_b": "electronic", + "choice_c": "punkrock", + "choice_d": "hiphop", + "answer_gt": "punkrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19964 + }, + { + "path": "76528.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "funk", + "choice_b": "chillout", + "choice_c": "electronic", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19965 + }, + { + "path": "1007839.mp3", + "question": "classify the genre of this music?", + "choice_a": "poprock", + "choice_b": "trance", + "choice_c": "hardrock", + "choice_d": "alternative", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19966 + }, + { + "path": "1182036.mp3", + "question": "detect the genre of this music?", + "choice_a": "electronic", + "choice_b": "popfolk", + "choice_c": "acidjazz", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19967 + }, + { + "path": "1415800.mp3", + "question": "specify the music genre here?", + "choice_a": "swing", + "choice_b": "electronic", + "choice_c": "folk", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19968 + }, + { + "path": "1149753.mp3", + "question": "specify the music genre here?", + "choice_a": "easylistening", + "choice_b": "classical", + "choice_c": "funk", + "choice_d": "reggae", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19969 + }, + { + "path": "1353768.mp3", + "question": "What style of music is this?", + "choice_a": "electronica", + "choice_b": "soundtrack", + "choice_c": "minimal", + "choice_d": "chillout", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19970 + }, + { + "path": "1026043.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "newage", + "choice_b": "jazz", + "choice_c": "techno", + "choice_d": "blues", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19971 + }, + { + "path": "44432.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "newwave", + "choice_b": "psychedelic", + "choice_c": "rock", + "choice_d": "folk", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19972 + }, + { + "path": "982408.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "soundtrack", + "choice_b": "instrumentalpop", + "choice_c": "progressive", + "choice_d": "indie", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19973 + }, + { + "path": "493294.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "darkambient", + "choice_b": "alternative", + "choice_c": "idm", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19974 + }, + { + "path": "1410639.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "industrial", + "choice_b": "psychedelic", + "choice_c": "breakbeat", + "choice_d": "house", + "answer_gt": "psychedelic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19975 + }, + { + "path": "1267744.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "rap", + "choice_b": "soundtrack", + "choice_c": "ethno", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19976 + }, + { + "path": "270103.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "indie", + "choice_b": "classical", + "choice_c": "popfolk", + "choice_d": "rnb", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19977 + }, + { + "path": "979849.mp3", + "question": "detect the genre of this music?", + "choice_a": "triphop", + "choice_b": "darkambient", + "choice_c": "soul", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19978 + }, + { + "path": "1124837.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "instrumentalpop", + "choice_b": "popfolk", + "choice_c": "easylistening", + "choice_d": "world", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19979 + }, + { + "path": "1026345.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "funk", + "choice_b": "electronic", + "choice_c": "rocknroll", + "choice_d": "easylistening", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19980 + }, + { + "path": "979847.mp3", + "question": "What style of music is this?", + "choice_a": "atmospheric", + "choice_b": "ambient", + "choice_c": "easylistening", + "choice_d": "rock", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19981 + }, + { + "path": "13291.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "jazz", + "choice_b": "swing", + "choice_c": "rock", + "choice_d": "rocknroll", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19982 + }, + { + "path": "1410638.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "alternative", + "choice_b": "rnb", + "choice_c": "psychedelic", + "choice_d": "orchestral", + "answer_gt": "psychedelic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19983 + }, + { + "path": "179541.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "classical", + "choice_b": "acidjazz", + "choice_c": "hiphop", + "choice_d": "poprock", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19984 + }, + { + "path": "953631.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "popfolk", + "choice_b": "funk", + "choice_c": "rock", + "choice_d": "poprock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19985 + }, + { + "path": "133675.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "indie", + "choice_b": "electronic", + "choice_c": "newwave", + "choice_d": "breakbeat", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19986 + }, + { + "path": "1408400.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "darkambient", + "choice_b": "jazz", + "choice_c": "trance", + "choice_d": "ethno", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19987 + }, + { + "path": "1393142.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "psychedelic", + "choice_b": "easylistening", + "choice_c": "bossanova", + "choice_d": "breakbeat", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19988 + }, + { + "path": "914765.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "folk", + "choice_b": "latin", + "choice_c": "classical", + "choice_d": "hard", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19989 + }, + { + "path": "147668.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electropop", + "choice_b": "drumnbass", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19990 + }, + { + "path": "1234577.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "popfolk", + "choice_b": "easylistening", + "choice_c": "trance", + "choice_d": "poprock", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19991 + }, + { + "path": "1168491.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "minimal", + "choice_b": "techno", + "choice_c": "electronic", + "choice_d": "atmospheric", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19992 + }, + { + "path": "912119.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "experimental", + "choice_b": "dance", + "choice_c": "electronic", + "choice_d": "instrumentalpop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19993 + }, + { + "path": "1031402.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "idm", + "choice_b": "latin", + "choice_c": "triphop", + "choice_d": "folk", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19994 + }, + { + "path": "1187036.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "soundtrack", + "choice_b": "psychedelic", + "choice_c": "ambient", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19995 + }, + { + "path": "127864.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "soundtrack", + "choice_b": "acidjazz", + "choice_c": "ambient", + "choice_d": "metal", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19996 + }, + { + "path": "1351652.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "breakbeat", + "choice_b": "experimental", + "choice_c": "world", + "choice_d": "indie", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19997 + }, + { + "path": "208297.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "rock", + "choice_b": "poprock", + "choice_c": "triphop", + "choice_d": "electropop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19998 + }, + { + "path": "1187047.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "rap", + "choice_b": "breakbeat", + "choice_c": "triphop", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 19999 + }, + { + "path": "896191.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "newage", + "choice_b": "newwave", + "choice_c": "pop", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20000 + }, + { + "path": "1354585.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "darkambient", + "choice_b": "soundtrack", + "choice_c": "latin", + "choice_d": "psychedelic", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20001 + }, + { + "path": "376240.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "metal", + "choice_b": "swing", + "choice_c": "orchestral", + "choice_d": "techno", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20002 + }, + { + "path": "191401.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "pop", + "choice_b": "progressive", + "choice_c": "industrial", + "choice_d": "minimal", + "answer_gt": "industrial", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20003 + }, + { + "path": "1197948.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "metal", + "choice_b": "soul", + "choice_c": "ambient", + "choice_d": "alternative", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20004 + }, + { + "path": "985180.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "celtic", + "choice_b": "house", + "choice_c": "electronic", + "choice_d": "classical", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20005 + }, + { + "path": "1088586.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "hiphop", + "choice_b": "trance", + "choice_c": "pop", + "choice_d": "ambient", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20006 + }, + { + "path": "484443.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "celtic", + "choice_b": "indie", + "choice_c": "rnb", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20007 + }, + { + "path": "1133075.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "dance", + "choice_b": "world", + "choice_c": "idm", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20008 + }, + { + "path": "1352138.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "rocknroll", + "choice_b": "newwave", + "choice_c": "soundtrack", + "choice_d": "hiphop", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20009 + }, + { + "path": "15066.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "electronic", + "choice_b": "poprock", + "choice_c": "easylistening", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20010 + }, + { + "path": "156965.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "hiphop", + "choice_b": "instrumentalpop", + "choice_c": "ethno", + "choice_d": "breakbeat", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20011 + }, + { + "path": "1350990.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "swing", + "choice_b": "soundtrack", + "choice_c": "world", + "choice_d": "punkrock", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20012 + }, + { + "path": "1087538.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "hardrock", + "choice_b": "acidjazz", + "choice_c": "blues", + "choice_d": "bossanova", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20013 + }, + { + "path": "173521.mp3", + "question": "What style of music is this?", + "choice_a": "dance", + "choice_b": "electronic", + "choice_c": "indie", + "choice_d": "orchestral", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20014 + }, + { + "path": "1350948.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "popfolk", + "choice_b": "acidjazz", + "choice_c": "soundtrack", + "choice_d": "alternative", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20015 + }, + { + "path": "31493.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "pop", + "choice_b": "newage", + "choice_c": "hardrock", + "choice_d": "orchestral", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20016 + }, + { + "path": "799909.mp3", + "question": "What style of music is this?", + "choice_a": "easylistening", + "choice_b": "funk", + "choice_c": "trance", + "choice_d": "chanson", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20017 + }, + { + "path": "812119.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "hard", + "choice_b": "rnb", + "choice_c": "soundtrack", + "choice_d": "industrial", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20018 + }, + { + "path": "1236360.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "trance", + "choice_b": "darkambient", + "choice_c": "electropop", + "choice_d": "soul", + "answer_gt": "electropop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20019 + }, + { + "path": "23197.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "celtic", + "choice_b": "alternative", + "choice_c": "pop", + "choice_d": "darkambient", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20020 + }, + { + "path": "1099433.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "swing", + "choice_b": "ambient", + "choice_c": "pop", + "choice_d": "metal", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20021 + }, + { + "path": "458587.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "atmospheric", + "choice_b": "chanson", + "choice_c": "blues", + "choice_d": "trance", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20022 + }, + { + "path": "1286483.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "newage", + "choice_b": "folk", + "choice_c": "industrial", + "choice_d": "electronic", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20023 + }, + { + "path": "1051721.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "blues", + "choice_b": "rock", + "choice_c": "electropop", + "choice_d": "latin", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20024 + }, + { + "path": "1164047.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "chillout", + "choice_b": "punkrock", + "choice_c": "folk", + "choice_d": "poprock", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20025 + }, + { + "path": "431450.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "classical", + "choice_b": "poprock", + "choice_c": "rocknroll", + "choice_d": "psychedelic", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20026 + }, + { + "path": "147714.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "newage", + "choice_b": "country", + "choice_c": "classical", + "choice_d": "hardrock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20027 + }, + { + "path": "1353767.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "indie", + "choice_b": "blues", + "choice_c": "soundtrack", + "choice_d": "minimal", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20028 + }, + { + "path": "178087.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "breakbeat", + "choice_b": "classical", + "choice_c": "indie", + "choice_d": "metal", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20029 + }, + { + "path": "199879.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "bossanova", + "choice_b": "electronic", + "choice_c": "progressive", + "choice_d": "experimental", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20030 + }, + { + "path": "588703.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "celtic", + "choice_b": "electronic", + "choice_c": "house", + "choice_d": "minimal", + "answer_gt": "minimal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20031 + }, + { + "path": "1026822.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "swing", + "choice_b": "trance", + "choice_c": "latin", + "choice_d": "popfolk", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20032 + }, + { + "path": "1089273.mp3", + "question": "what's the genre of this music?", + "choice_a": "chanson", + "choice_b": "classical", + "choice_c": "techno", + "choice_d": "hiphop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20033 + }, + { + "path": "1085244.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "minimal", + "choice_b": "metal", + "choice_c": "classical", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20034 + }, + { + "path": "172817.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "acidjazz", + "choice_b": "electronic", + "choice_c": "instrumentalpop", + "choice_d": "country", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20035 + }, + { + "path": "322123.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "house", + "choice_b": "soul", + "choice_c": "rocknroll", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20036 + }, + { + "path": "90119.mp3", + "question": "detect the genre of this music?", + "choice_a": "reggae", + "choice_b": "blues", + "choice_c": "experimental", + "choice_d": "industrial", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20037 + }, + { + "path": "1076570.mp3", + "question": "classify the genre of this music?", + "choice_a": "jazz", + "choice_b": "house", + "choice_c": "blues", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20038 + }, + { + "path": "1321120.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "hard", + "choice_b": "world", + "choice_c": "classical", + "choice_d": "dance", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20039 + }, + { + "path": "130313.mp3", + "question": "What's the classification of this music?", + "choice_a": "hardrock", + "choice_b": "blues", + "choice_c": "electronica", + "choice_d": "experimental", + "answer_gt": "hardrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20040 + }, + { + "path": "985168.mp3", + "question": "What's the classification of this music?", + "choice_a": "hiphop", + "choice_b": "latin", + "choice_c": "idm", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20041 + }, + { + "path": "371453.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "reggae", + "choice_b": "electronic", + "choice_c": "pop", + "choice_d": "electronica", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20042 + }, + { + "path": "842780.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "trance", + "choice_b": "soul", + "choice_c": "world", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20043 + }, + { + "path": "20492.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "jazz", + "choice_b": "triphop", + "choice_c": "chillout", + "choice_d": "hard", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20044 + }, + { + "path": "1086601.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "minimal", + "choice_b": "hiphop", + "choice_c": "world", + "choice_d": "electropop", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20045 + }, + { + "path": "253254.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "drumnbass", + "choice_b": "psychedelic", + "choice_c": "ambient", + "choice_d": "chanson", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20046 + }, + { + "path": "2966.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "classical", + "choice_b": "psychedelic", + "choice_c": "chanson", + "choice_d": "rock", + "answer_gt": "chanson", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20047 + }, + { + "path": "1121565.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "pop", + "choice_b": "alternative", + "choice_c": "breakbeat", + "choice_d": "funk", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20048 + }, + { + "path": "683122.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "alternative", + "choice_b": "hard", + "choice_c": "ambient", + "choice_d": "metal", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20049 + }, + { + "path": "1121566.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "latin", + "choice_b": "alternative", + "choice_c": "poprock", + "choice_d": "swing", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20050 + }, + { + "path": "1086643.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "funk", + "choice_b": "blues", + "choice_c": "folk", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20051 + }, + { + "path": "1046703.mp3", + "question": "what's the genre of this music?", + "choice_a": "newage", + "choice_b": "techno", + "choice_c": "house", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20052 + }, + { + "path": "13289.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "orchestral", + "choice_b": "rock", + "choice_c": "poprock", + "choice_d": "easylistening", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20053 + }, + { + "path": "1083531.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "pop", + "choice_b": "trance", + "choice_c": "celtic", + "choice_d": "triphop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20054 + }, + { + "path": "814949.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "progressive", + "choice_b": "triphop", + "choice_c": "acidjazz", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20055 + }, + { + "path": "184242.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "instrumentalpop", + "choice_b": "ambient", + "choice_c": "industrial", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20056 + }, + { + "path": "208187.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "rnb", + "choice_b": "rock", + "choice_c": "classical", + "choice_d": "popfolk", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20057 + }, + { + "path": "76949.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "darkambient", + "choice_b": "electronic", + "choice_c": "pop", + "choice_d": "instrumentalpop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20058 + }, + { + "path": "1160224.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "breakbeat", + "choice_b": "techno", + "choice_c": "folk", + "choice_d": "rocknroll", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20059 + }, + { + "path": "288374.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "classical", + "choice_b": "darkambient", + "choice_c": "rocknroll", + "choice_d": "rnb", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20060 + }, + { + "path": "379520.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "hiphop", + "choice_b": "techno", + "choice_c": "atmospheric", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20061 + }, + { + "path": "78361.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "electronica", + "choice_b": "dance", + "choice_c": "bossanova", + "choice_d": "breakbeat", + "answer_gt": "bossanova", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20062 + }, + { + "path": "1286491.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "dance", + "choice_b": "breakbeat", + "choice_c": "newage", + "choice_d": "metal", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20063 + }, + { + "path": "1128252.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "triphop", + "choice_b": "blues", + "choice_c": "swing", + "choice_d": "celtic", + "answer_gt": "celtic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20064 + }, + { + "path": "121310.mp3", + "question": "what's the genre of this music?", + "choice_a": "progressive", + "choice_b": "rock", + "choice_c": "electronic", + "choice_d": "dance", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20065 + }, + { + "path": "549222.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "easylistening", + "choice_b": "indie", + "choice_c": "celtic", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20066 + }, + { + "path": "847200.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "soundtrack", + "choice_b": "classical", + "choice_c": "world", + "choice_d": "hardrock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20067 + }, + { + "path": "1062498.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "latin", + "choice_b": "ambient", + "choice_c": "reggae", + "choice_d": "popfolk", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20068 + }, + { + "path": "1127781.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "hiphop", + "choice_b": "popfolk", + "choice_c": "indie", + "choice_d": "soundtrack", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20069 + }, + { + "path": "1227497.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "instrumentalpop", + "choice_b": "soundtrack", + "choice_c": "easylistening", + "choice_d": "chanson", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20070 + }, + { + "path": "925537.mp3", + "question": "What's the classification of this music?", + "choice_a": "rocknroll", + "choice_b": "punkrock", + "choice_c": "poprock", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20071 + }, + { + "path": "240537.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "country", + "choice_b": "dance", + "choice_c": "trance", + "choice_d": "idm", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20072 + }, + { + "path": "1352581.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "trance", + "choice_b": "hard", + "choice_c": "soundtrack", + "choice_d": "acidjazz", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20073 + }, + { + "path": "431446.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "classical", + "choice_b": "house", + "choice_c": "metal", + "choice_d": "swing", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20074 + }, + { + "path": "376248.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "world", + "choice_b": "hard", + "choice_c": "metal", + "choice_d": "funk", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20075 + }, + { + "path": "324381.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "techno", + "choice_b": "breakbeat", + "choice_c": "easylistening", + "choice_d": "funk", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20076 + }, + { + "path": "447853.mp3", + "question": "What's the classification of this music?", + "choice_a": "rock", + "choice_b": "house", + "choice_c": "easylistening", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20077 + }, + { + "path": "1167076.mp3", + "question": "What style of music is this?", + "choice_a": "soul", + "choice_b": "classical", + "choice_c": "punkrock", + "choice_d": "popfolk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20078 + }, + { + "path": "345621.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "psychedelic", + "choice_b": "folk", + "choice_c": "breakbeat", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20079 + }, + { + "path": "3731.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "electronic", + "choice_b": "classical", + "choice_c": "reggae", + "choice_d": "acidjazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20080 + }, + { + "path": "814943.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "reggae", + "choice_b": "chillout", + "choice_c": "bossanova", + "choice_d": "techno", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20081 + }, + { + "path": "76956.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "darkambient", + "choice_b": "pop", + "choice_c": "orchestral", + "choice_d": "newwave", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20082 + }, + { + "path": "285409.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "reggae", + "choice_b": "hard", + "choice_c": "electronic", + "choice_d": "ambient", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20083 + }, + { + "path": "923463.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "jazz", + "choice_b": "darkambient", + "choice_c": "electronica", + "choice_d": "country", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20084 + }, + { + "path": "502330.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "latin", + "choice_b": "swing", + "choice_c": "classical", + "choice_d": "soundtrack", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20085 + }, + { + "path": "1126025.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "hiphop", + "choice_b": "celtic", + "choice_c": "psychedelic", + "choice_d": "rocknroll", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20086 + }, + { + "path": "1374283.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "hiphop", + "choice_b": "soul", + "choice_c": "swing", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20087 + }, + { + "path": "858545.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "electronic", + "choice_b": "idm", + "choice_c": "indie", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20088 + }, + { + "path": "1344148.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "orchestral", + "choice_b": "rock", + "choice_c": "hard", + "choice_d": "newwave", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20089 + }, + { + "path": "1372930.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "rock", + "choice_b": "trance", + "choice_c": "ambient", + "choice_d": "darkambient", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20090 + }, + { + "path": "1050999.mp3", + "question": "identify the music genre of this track.", + "choice_a": "indie", + "choice_b": "easylistening", + "choice_c": "electronic", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20091 + }, + { + "path": "1377207.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "reggae", + "choice_b": "celtic", + "choice_c": "soul", + "choice_d": "folk", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20092 + }, + { + "path": "90125.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "reggae", + "choice_b": "punkrock", + "choice_c": "newwave", + "choice_d": "rap", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20093 + }, + { + "path": "283379.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "industrial", + "choice_b": "electronic", + "choice_c": "techno", + "choice_d": "progressive", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20094 + }, + { + "path": "1067228.mp3", + "question": "what's the genre of this music?", + "choice_a": "minimal", + "choice_b": "electronic", + "choice_c": "breakbeat", + "choice_d": "orchestral", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20095 + }, + { + "path": "393247.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "idm", + "choice_b": "classical", + "choice_c": "pop", + "choice_d": "progressive", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20096 + }, + { + "path": "1351053.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "ethno", + "choice_b": "metal", + "choice_c": "trance", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20097 + }, + { + "path": "434897.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "punkrock", + "choice_b": "rnb", + "choice_c": "swing", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20098 + }, + { + "path": "749823.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "chanson", + "choice_b": "celtic", + "choice_c": "ambient", + "choice_d": "rock", + "answer_gt": "chanson", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20099 + }, + { + "path": "979856.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "ambient", + "choice_b": "swing", + "choice_c": "dance", + "choice_d": "poprock", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20100 + }, + { + "path": "965450.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "soul", + "choice_b": "hiphop", + "choice_c": "instrumentalpop", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20101 + }, + { + "path": "211971.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "alternative", + "choice_b": "experimental", + "choice_c": "orchestral", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20102 + }, + { + "path": "1126859.mp3", + "question": "What style of music is this?", + "choice_a": "hiphop", + "choice_b": "drumnbass", + "choice_c": "electronic", + "choice_d": "ethno", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20103 + }, + { + "path": "127546.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "country", + "choice_b": "ethno", + "choice_c": "electropop", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20104 + }, + { + "path": "1106894.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "ethno", + "choice_b": "industrial", + "choice_c": "ambient", + "choice_d": "trance", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20105 + }, + { + "path": "147292.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "world", + "choice_b": "electronica", + "choice_c": "electronic", + "choice_d": "swing", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20106 + }, + { + "path": "1088590.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "pop", + "choice_b": "hardrock", + "choice_c": "country", + "choice_d": "minimal", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20107 + }, + { + "path": "1085247.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "soul", + "choice_b": "swing", + "choice_c": "classical", + "choice_d": "ambient", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20108 + }, + { + "path": "285408.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "electronic", + "choice_b": "electropop", + "choice_c": "ethno", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20109 + }, + { + "path": "1419773.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electronic", + "choice_b": "soundtrack", + "choice_c": "jazz", + "choice_d": "techno", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20110 + }, + { + "path": "23008.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "triphop", + "choice_b": "darkambient", + "choice_c": "orchestral", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20111 + }, + { + "path": "1105211.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "bossanova", + "choice_b": "jazz", + "choice_c": "instrumentalpop", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20112 + }, + { + "path": "1297357.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "metal", + "choice_b": "soundtrack", + "choice_c": "triphop", + "choice_d": "reggae", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20113 + }, + { + "path": "223590.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "funk", + "choice_b": "electronica", + "choice_c": "ambient", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20114 + }, + { + "path": "508152.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "rock", + "choice_b": "triphop", + "choice_c": "punkrock", + "choice_d": "breakbeat", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20115 + }, + { + "path": "1120127.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "reggae", + "choice_b": "experimental", + "choice_c": "dance", + "choice_d": "chillout", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20116 + }, + { + "path": "1109774.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "orchestral", + "choice_b": "country", + "choice_c": "electronic", + "choice_d": "hardrock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20117 + }, + { + "path": "13288.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "rock", + "choice_b": "country", + "choice_c": "blues", + "choice_d": "electronic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20118 + }, + { + "path": "1125352.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "electronic", + "choice_b": "rocknroll", + "choice_c": "poprock", + "choice_d": "industrial", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20119 + }, + { + "path": "353652.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "latin", + "choice_b": "industrial", + "choice_c": "electronic", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20120 + }, + { + "path": "221665.mp3", + "question": "detect the genre of this music?", + "choice_a": "house", + "choice_b": "rock", + "choice_c": "psychedelic", + "choice_d": "pop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20121 + }, + { + "path": "1046641.mp3", + "question": "What style of music is this?", + "choice_a": "orchestral", + "choice_b": "ethno", + "choice_c": "rock", + "choice_d": "folk", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20122 + }, + { + "path": "1354787.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "poprock", + "choice_b": "rock", + "choice_c": "jazz", + "choice_d": "punkrock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20123 + }, + { + "path": "1127775.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "reggae", + "choice_b": "poprock", + "choice_c": "darkambient", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20124 + }, + { + "path": "147667.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "experimental", + "choice_b": "country", + "choice_c": "electronic", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20125 + }, + { + "path": "172816.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "rock", + "choice_b": "experimental", + "choice_c": "electronic", + "choice_d": "trance", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20126 + }, + { + "path": "1301831.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "pop", + "choice_b": "triphop", + "choice_c": "hard", + "choice_d": "rnb", + "answer_gt": "rnb", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20127 + }, + { + "path": "147712.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "techno", + "choice_b": "soul", + "choice_c": "classical", + "choice_d": "electronica", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20128 + }, + { + "path": "1070637.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "classical", + "choice_b": "idm", + "choice_c": "newwave", + "choice_d": "instrumentalpop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20129 + }, + { + "path": "1362267.mp3", + "question": "specify the music genre here?", + "choice_a": "atmospheric", + "choice_b": "easylistening", + "choice_c": "techno", + "choice_d": "triphop", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20130 + }, + { + "path": "1150771.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "rnb", + "choice_b": "breakbeat", + "choice_c": "techno", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20131 + }, + { + "path": "16092.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "metal", + "choice_b": "ethno", + "choice_c": "electronica", + "choice_d": "celtic", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20132 + }, + { + "path": "1419657.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "classical", + "choice_b": "hardrock", + "choice_c": "latin", + "choice_d": "electropop", + "answer_gt": "hardrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20133 + }, + { + "path": "1070638.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "hard", + "choice_b": "classical", + "choice_c": "rock", + "choice_d": "minimal", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20134 + }, + { + "path": "1286518.mp3", + "question": "What style of music is this?", + "choice_a": "newage", + "choice_b": "darkambient", + "choice_c": "breakbeat", + "choice_d": "techno", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20135 + }, + { + "path": "1246808.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "newwave", + "choice_b": "indie", + "choice_c": "triphop", + "choice_d": "idm", + "answer_gt": "newwave", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20136 + }, + { + "path": "431431.mp3", + "question": "detect the genre of this music?", + "choice_a": "minimal", + "choice_b": "indie", + "choice_c": "classical", + "choice_d": "alternative", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20137 + }, + { + "path": "1286517.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "easylistening", + "choice_b": "newage", + "choice_c": "popfolk", + "choice_d": "country", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20138 + }, + { + "path": "1276485.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "pop", + "choice_b": "blues", + "choice_c": "minimal", + "choice_d": "techno", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20139 + }, + { + "path": "979842.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "ambient", + "choice_b": "country", + "choice_c": "progressive", + "choice_d": "rock", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20140 + }, + { + "path": "1068780.mp3", + "question": "what's the genre of this music?", + "choice_a": "acidjazz", + "choice_b": "pop", + "choice_c": "instrumentalpop", + "choice_d": "rnb", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20141 + }, + { + "path": "191411.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "idm", + "choice_b": "alternative", + "choice_c": "industrial", + "choice_d": "chanson", + "answer_gt": "industrial", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20142 + }, + { + "path": "549228.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "hard", + "choice_b": "indie", + "choice_c": "darkambient", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20143 + }, + { + "path": "434862.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "drumnbass", + "choice_b": "pop", + "choice_c": "hard", + "choice_d": "jazz", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20144 + }, + { + "path": "1350308.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "chanson", + "choice_b": "orchestral", + "choice_c": "soundtrack", + "choice_d": "ethno", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20145 + }, + { + "path": "290003.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electronic", + "choice_b": "techno", + "choice_c": "breakbeat", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20146 + }, + { + "path": "344330.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "electronic", + "choice_b": "country", + "choice_c": "soundtrack", + "choice_d": "bossanova", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20147 + }, + { + "path": "493297.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "pop", + "choice_b": "country", + "choice_c": "newwave", + "choice_d": "easylistening", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20148 + }, + { + "path": "1270497.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "electronica", + "choice_b": "ambient", + "choice_c": "industrial", + "choice_d": "latin", + "answer_gt": "electronica", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20149 + }, + { + "path": "1332223.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "rap", + "choice_b": "rnb", + "choice_c": "classical", + "choice_d": "atmospheric", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20150 + }, + { + "path": "135662.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "hiphop", + "choice_b": "punkrock", + "choice_c": "easylistening", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20151 + }, + { + "path": "1372913.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "triphop", + "choice_b": "rap", + "choice_c": "easylistening", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20152 + }, + { + "path": "172018.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "rap", + "choice_b": "funk", + "choice_c": "reggae", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20153 + }, + { + "path": "1415797.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "atmospheric", + "choice_b": "celtic", + "choice_c": "idm", + "choice_d": "alternative", + "answer_gt": "atmospheric", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20154 + }, + { + "path": "383978.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "latin", + "choice_b": "newwave", + "choice_c": "alternative", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20155 + }, + { + "path": "208283.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "progressive", + "choice_b": "metal", + "choice_c": "country", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20156 + }, + { + "path": "1041285.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "hiphop", + "choice_b": "soundtrack", + "choice_c": "ambient", + "choice_d": "hard", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20157 + }, + { + "path": "345618.mp3", + "question": "specify the music genre here?", + "choice_a": "rock", + "choice_b": "idm", + "choice_c": "hard", + "choice_d": "electronic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20158 + }, + { + "path": "243824.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "trance", + "choice_b": "classical", + "choice_c": "rock", + "choice_d": "idm", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20159 + }, + { + "path": "1126848.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "experimental", + "choice_b": "classical", + "choice_c": "alternative", + "choice_d": "darkambient", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20160 + }, + { + "path": "276972.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "easylistening", + "choice_b": "dance", + "choice_c": "experimental", + "choice_d": "hard", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20161 + }, + { + "path": "1331776.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "chillout", + "choice_b": "newage", + "choice_c": "breakbeat", + "choice_d": "chanson", + "answer_gt": "breakbeat", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20162 + }, + { + "path": "149072.mp3", + "question": "detect the genre of this music?", + "choice_a": "rock", + "choice_b": "world", + "choice_c": "minimal", + "choice_d": "ethno", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20163 + }, + { + "path": "680851.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "orchestral", + "choice_b": "alternative", + "choice_c": "rock", + "choice_d": "electronica", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20164 + }, + { + "path": "1086681.mp3", + "question": "What style of music is this?", + "choice_a": "folk", + "choice_b": "trance", + "choice_c": "classical", + "choice_d": "funk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20165 + }, + { + "path": "842725.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "house", + "choice_b": "hiphop", + "choice_c": "dance", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20166 + }, + { + "path": "1126855.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "hiphop", + "choice_b": "hard", + "choice_c": "electronica", + "choice_d": "orchestral", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20167 + }, + { + "path": "1149755.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "drumnbass", + "choice_b": "classical", + "choice_c": "alternative", + "choice_d": "industrial", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20168 + }, + { + "path": "368174.mp3", + "question": "detect the genre of this music?", + "choice_a": "orchestral", + "choice_b": "classical", + "choice_c": "progressive", + "choice_d": "instrumentalpop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20169 + }, + { + "path": "1398527.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "soul", + "choice_b": "minimal", + "choice_c": "pop", + "choice_d": "rap", + "answer_gt": "rap", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20170 + }, + { + "path": "283821.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "hard", + "choice_b": "punkrock", + "choice_c": "electronic", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20171 + }, + { + "path": "253751.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "easylistening", + "choice_b": "reggae", + "choice_c": "breakbeat", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20172 + }, + { + "path": "616362.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "acidjazz", + "choice_b": "soul", + "choice_c": "jazz", + "choice_d": "electronic", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20173 + }, + { + "path": "979841.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "metal", + "choice_b": "ambient", + "choice_c": "country", + "choice_d": "rocknroll", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20174 + }, + { + "path": "32786.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "rock", + "choice_b": "punkrock", + "choice_c": "classical", + "choice_d": "dance", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20175 + }, + { + "path": "847337.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "electronica", + "choice_b": "electronic", + "choice_c": "reggae", + "choice_d": "newwave", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20176 + }, + { + "path": "979871.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "trance", + "choice_b": "country", + "choice_c": "ambient", + "choice_d": "pop", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20177 + }, + { + "path": "23009.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electronic", + "choice_b": "soundtrack", + "choice_c": "bossanova", + "choice_d": "drumnbass", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20178 + }, + { + "path": "371455.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "hard", + "choice_b": "punkrock", + "choice_c": "soul", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20179 + }, + { + "path": "49235.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "trance", + "choice_b": "rock", + "choice_c": "orchestral", + "choice_d": "newage", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20180 + }, + { + "path": "1159862.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "reggae", + "choice_b": "alternative", + "choice_c": "drumnbass", + "choice_d": "indie", + "answer_gt": "indie", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20181 + }, + { + "path": "431430.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "hardrock", + "choice_b": "ethno", + "choice_c": "drumnbass", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20182 + }, + { + "path": "15067.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "instrumentalpop", + "choice_b": "darkambient", + "choice_c": "jazz", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20183 + }, + { + "path": "972771.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "electronica", + "choice_b": "electronic", + "choice_c": "progressive", + "choice_d": "blues", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20184 + }, + { + "path": "1029179.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "ambient", + "choice_b": "chillout", + "choice_c": "house", + "choice_d": "electronic", + "answer_gt": "chillout", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20185 + }, + { + "path": "1354580.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "dance", + "choice_b": "soundtrack", + "choice_c": "indie", + "choice_d": "chanson", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20186 + }, + { + "path": "1096088.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "rock", + "choice_b": "progressive", + "choice_c": "ambient", + "choice_d": "orchestral", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20187 + }, + { + "path": "253900.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "electronic", + "choice_b": "bossanova", + "choice_c": "easylistening", + "choice_d": "trance", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20188 + }, + { + "path": "966520.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "soul", + "choice_b": "world", + "choice_c": "soundtrack", + "choice_d": "bossanova", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20189 + }, + { + "path": "639279.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "ambient", + "choice_b": "idm", + "choice_c": "experimental", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20190 + }, + { + "path": "1159853.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "jazz", + "choice_b": "dance", + "choice_c": "electronic", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20191 + }, + { + "path": "1330560.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "world", + "choice_b": "chanson", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20192 + }, + { + "path": "1286516.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "newage", + "choice_b": "industrial", + "choice_c": "drumnbass", + "choice_d": "swing", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20193 + }, + { + "path": "1197942.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "classical", + "choice_b": "metal", + "choice_c": "poprock", + "choice_d": "drumnbass", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20194 + }, + { + "path": "1166261.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "soundtrack", + "choice_b": "darkambient", + "choice_c": "psychedelic", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20195 + }, + { + "path": "1279179.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "ambient", + "choice_b": "techno", + "choice_c": "rnb", + "choice_d": "world", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20196 + }, + { + "path": "240495.mp3", + "question": "What style of music is this?", + "choice_a": "atmospheric", + "choice_b": "dance", + "choice_c": "latin", + "choice_d": "psychedelic", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20197 + }, + { + "path": "110665.mp3", + "question": "identify the music genre of this track.", + "choice_a": "newwave", + "choice_b": "soundtrack", + "choice_c": "punkrock", + "choice_d": "funk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20198 + }, + { + "path": "1128823.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "reggae", + "choice_b": "latin", + "choice_c": "soundtrack", + "choice_d": "darkambient", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20199 + }, + { + "path": "251707.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "world", + "choice_b": "drumnbass", + "choice_c": "easylistening", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20200 + }, + { + "path": "202647.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "breakbeat", + "choice_b": "soul", + "choice_c": "darkambient", + "choice_d": "jazz", + "answer_gt": "soul", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20201 + }, + { + "path": "879150.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "dance", + "choice_b": "electropop", + "choice_c": "jazz", + "choice_d": "chanson", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20202 + }, + { + "path": "458675.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "progressive", + "choice_b": "hiphop", + "choice_c": "orchestral", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20203 + }, + { + "path": "812117.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "rap", + "choice_b": "idm", + "choice_c": "country", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20204 + }, + { + "path": "1214583.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "psychedelic", + "choice_b": "darkambient", + "choice_c": "hardrock", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20205 + }, + { + "path": "76953.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "pop", + "choice_b": "rap", + "choice_c": "newwave", + "choice_d": "reggae", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20206 + }, + { + "path": "956428.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "industrial", + "choice_b": "progressive", + "choice_c": "indie", + "choice_d": "metal", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20207 + }, + { + "path": "246737.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "darkambient", + "choice_b": "electronica", + "choice_c": "acidjazz", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20208 + }, + { + "path": "1389036.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "trance", + "choice_b": "metal", + "choice_c": "swing", + "choice_d": "classical", + "answer_gt": "swing", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20209 + }, + { + "path": "1354146.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "triphop", + "choice_b": "folk", + "choice_c": "soundtrack", + "choice_d": "rocknroll", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20210 + }, + { + "path": "156963.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "electronica", + "choice_b": "swing", + "choice_c": "soundtrack", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20211 + }, + { + "path": "896193.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "hardrock", + "choice_b": "electronic", + "choice_c": "drumnbass", + "choice_d": "industrial", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20212 + }, + { + "path": "199209.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "rnb", + "choice_b": "folk", + "choice_c": "chanson", + "choice_d": "acidjazz", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20213 + }, + { + "path": "1419659.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "newwave", + "choice_b": "country", + "choice_c": "hardrock", + "choice_d": "instrumentalpop", + "answer_gt": "hardrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20214 + }, + { + "path": "13294.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "rock", + "choice_b": "progressive", + "choice_c": "minimal", + "choice_d": "celtic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20215 + }, + { + "path": "842757.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "idm", + "choice_b": "celtic", + "choice_c": "instrumentalpop", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20216 + }, + { + "path": "424247.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "classical", + "choice_b": "rap", + "choice_c": "techno", + "choice_d": "triphop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20217 + }, + { + "path": "1257400.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "reggae", + "choice_b": "poprock", + "choice_c": "country", + "choice_d": "ambient", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20218 + }, + { + "path": "202554.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "electronic", + "choice_b": "rock", + "choice_c": "atmospheric", + "choice_d": "rnb", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20219 + }, + { + "path": "1347951.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "reggae", + "choice_b": "orchestral", + "choice_c": "electronic", + "choice_d": "industrial", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20220 + }, + { + "path": "1114849.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "hard", + "choice_b": "electronica", + "choice_c": "pop", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20221 + }, + { + "path": "347199.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "house", + "choice_b": "triphop", + "choice_c": "rock", + "choice_d": "country", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20222 + }, + { + "path": "83399.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "latin", + "choice_b": "dance", + "choice_c": "blues", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20223 + }, + { + "path": "1349977.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "atmospheric", + "choice_b": "rocknroll", + "choice_c": "punkrock", + "choice_d": "experimental", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20224 + }, + { + "path": "1128247.mp3", + "question": "specify the music genre here?", + "choice_a": "atmospheric", + "choice_b": "celtic", + "choice_c": "soundtrack", + "choice_d": "dance", + "answer_gt": "celtic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20225 + }, + { + "path": "842781.mp3", + "question": "detect the genre of this music?", + "choice_a": "experimental", + "choice_b": "pop", + "choice_c": "latin", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20226 + }, + { + "path": "14017.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "instrumentalpop", + "choice_b": "classical", + "choice_c": "newage", + "choice_d": "newwave", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20227 + }, + { + "path": "243758.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "classical", + "choice_b": "hardrock", + "choice_c": "triphop", + "choice_d": "chanson", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20228 + }, + { + "path": "961359.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "indie", + "choice_b": "rnb", + "choice_c": "world", + "choice_d": "popfolk", + "answer_gt": "indie", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20229 + }, + { + "path": "147715.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "classical", + "choice_b": "ambient", + "choice_c": "darkambient", + "choice_d": "atmospheric", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20230 + }, + { + "path": "1109581.mp3", + "question": "identify the music genre of this track.", + "choice_a": "blues", + "choice_b": "celtic", + "choice_c": "pop", + "choice_d": "newwave", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20231 + }, + { + "path": "127891.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "experimental", + "choice_b": "chillout", + "choice_c": "rnb", + "choice_d": "ethno", + "answer_gt": "chillout", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20232 + }, + { + "path": "349401.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "soundtrack", + "choice_b": "newage", + "choice_c": "atmospheric", + "choice_d": "chillout", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20233 + }, + { + "path": "1044826.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "triphop", + "choice_b": "chillout", + "choice_c": "easylistening", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20234 + }, + { + "path": "1037343.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "hard", + "choice_b": "hardrock", + "choice_c": "soundtrack", + "choice_d": "breakbeat", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20235 + }, + { + "path": "1051706.mp3", + "question": "What's the classification of this music?", + "choice_a": "ethno", + "choice_b": "rock", + "choice_c": "jazz", + "choice_d": "industrial", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20236 + }, + { + "path": "208300.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "rock", + "choice_b": "rnb", + "choice_c": "easylistening", + "choice_d": "classical", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20237 + }, + { + "path": "202648.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "electronic", + "choice_b": "rnb", + "choice_c": "swing", + "choice_d": "newage", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20238 + }, + { + "path": "379514.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "electronica", + "choice_b": "electronic", + "choice_c": "bossanova", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20239 + }, + { + "path": "458673.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "rocknroll", + "choice_b": "electronic", + "choice_c": "progressive", + "choice_d": "hard", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20240 + }, + { + "path": "243702.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "hiphop", + "choice_b": "progressive", + "choice_c": "popfolk", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20241 + }, + { + "path": "434903.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "drumnbass", + "choice_b": "hard", + "choice_c": "progressive", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20242 + }, + { + "path": "1227500.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "industrial", + "choice_b": "soundtrack", + "choice_c": "classical", + "choice_d": "easylistening", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20243 + }, + { + "path": "1354588.mp3", + "question": "What style of music is this?", + "choice_a": "chanson", + "choice_b": "house", + "choice_c": "soundtrack", + "choice_d": "soul", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20244 + }, + { + "path": "226683.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "rock", + "choice_b": "psychedelic", + "choice_c": "country", + "choice_d": "poprock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20245 + }, + { + "path": "222087.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "electronic", + "choice_b": "triphop", + "choice_c": "bossanova", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20246 + }, + { + "path": "1045229.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "electronic", + "choice_b": "pop", + "choice_c": "dance", + "choice_d": "easylistening", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20247 + }, + { + "path": "383977.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "punkrock", + "choice_b": "pop", + "choice_c": "bossanova", + "choice_d": "soundtrack", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20248 + }, + { + "path": "944152.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "orchestral", + "choice_b": "classical", + "choice_c": "poprock", + "choice_d": "house", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20249 + }, + { + "path": "283215.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "ambient", + "choice_b": "soul", + "choice_c": "newage", + "choice_d": "hiphop", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20250 + }, + { + "path": "1072137.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "alternative", + "choice_b": "acidjazz", + "choice_c": "jazz", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20251 + }, + { + "path": "285650.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "minimal", + "choice_b": "metal", + "choice_c": "classical", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20252 + }, + { + "path": "1153924.mp3", + "question": "What's the classification of this music?", + "choice_a": "pop", + "choice_b": "acidjazz", + "choice_c": "ethno", + "choice_d": "newage", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20253 + }, + { + "path": "1289.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "house", + "choice_b": "ambient", + "choice_c": "rock", + "choice_d": "alternative", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20254 + }, + { + "path": "1391960.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "atmospheric", + "choice_b": "newage", + "choice_c": "house", + "choice_d": "electronica", + "answer_gt": "atmospheric", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20255 + }, + { + "path": "1393665.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "house", + "choice_b": "triphop", + "choice_c": "instrumentalpop", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20256 + }, + { + "path": "664815.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "popfolk", + "choice_b": "classical", + "choice_c": "minimal", + "choice_d": "rocknroll", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20257 + }, + { + "path": "1350823.mp3", + "question": "detect the genre of this music?", + "choice_a": "bossanova", + "choice_b": "jazz", + "choice_c": "electronic", + "choice_d": "atmospheric", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20258 + }, + { + "path": "1126031.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "world", + "choice_b": "hiphop", + "choice_c": "acidjazz", + "choice_d": "triphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20259 + }, + { + "path": "1349532.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "chanson", + "choice_b": "rocknroll", + "choice_c": "soundtrack", + "choice_d": "indie", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20260 + }, + { + "path": "208493.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "psychedelic", + "choice_b": "rnb", + "choice_c": "rock", + "choice_d": "electronica", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20261 + }, + { + "path": "1389209.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "bossanova", + "choice_b": "electronica", + "choice_c": "classical", + "choice_d": "chillout", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20262 + }, + { + "path": "393256.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "classical", + "choice_b": "easylistening", + "choice_c": "ambient", + "choice_d": "poprock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20263 + }, + { + "path": "283382.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "world", + "choice_b": "minimal", + "choice_c": "instrumentalpop", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20264 + }, + { + "path": "119375.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "jazz", + "choice_b": "pop", + "choice_c": "experimental", + "choice_d": "minimal", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20265 + }, + { + "path": "1394327.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "funk", + "choice_b": "alternative", + "choice_c": "darkambient", + "choice_d": "rap", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20266 + }, + { + "path": "1267752.mp3", + "question": "what's the genre of this music?", + "choice_a": "blues", + "choice_b": "indie", + "choice_c": "classical", + "choice_d": "popfolk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20267 + }, + { + "path": "1362283.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "poprock", + "choice_b": "ethno", + "choice_c": "chanson", + "choice_d": "progressive", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20268 + }, + { + "path": "1393669.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "swing", + "choice_b": "blues", + "choice_c": "indie", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20269 + }, + { + "path": "965448.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "hardrock", + "choice_b": "reggae", + "choice_c": "classical", + "choice_d": "popfolk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20270 + }, + { + "path": "211984.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "darkambient", + "choice_b": "rnb", + "choice_c": "electronic", + "choice_d": "house", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20271 + }, + { + "path": "951321.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "drumnbass", + "choice_b": "hiphop", + "choice_c": "instrumentalpop", + "choice_d": "latin", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20272 + }, + { + "path": "953635.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "easylistening", + "choice_b": "idm", + "choice_c": "newwave", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20273 + }, + { + "path": "371456.mp3", + "question": "what's the genre of this music?", + "choice_a": "swing", + "choice_b": "newwave", + "choice_c": "electronic", + "choice_d": "darkambient", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20274 + }, + { + "path": "944146.mp3", + "question": "what's the genre of this music?", + "choice_a": "poprock", + "choice_b": "pop", + "choice_c": "experimental", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20275 + }, + { + "path": "914767.mp3", + "question": "identify the music genre of this track.", + "choice_a": "techno", + "choice_b": "darkambient", + "choice_c": "electronic", + "choice_d": "folk", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20276 + }, + { + "path": "979862.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "soundtrack", + "choice_b": "classical", + "choice_c": "house", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20277 + }, + { + "path": "1096333.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "ambient", + "choice_b": "easylistening", + "choice_c": "swing", + "choice_d": "techno", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20278 + }, + { + "path": "1350544.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "metal", + "choice_b": "experimental", + "choice_c": "pop", + "choice_d": "swing", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20279 + }, + { + "path": "976716.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "rap", + "choice_b": "instrumentalpop", + "choice_c": "classical", + "choice_d": "techno", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20280 + }, + { + "path": "1121554.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "newage", + "choice_b": "blues", + "choice_c": "world", + "choice_d": "rocknroll", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20281 + }, + { + "path": "246745.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "bossanova", + "choice_b": "electronic", + "choice_c": "newage", + "choice_d": "instrumentalpop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20282 + }, + { + "path": "914769.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "folk", + "choice_b": "industrial", + "choice_c": "newage", + "choice_d": "poprock", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20283 + }, + { + "path": "965311.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "electronica", + "choice_b": "popfolk", + "choice_c": "classical", + "choice_d": "folk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20284 + }, + { + "path": "129577.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "ethno", + "choice_b": "newage", + "choice_c": "psychedelic", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20285 + }, + { + "path": "532906.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "atmospheric", + "choice_b": "pop", + "choice_c": "electronic", + "choice_d": "electronica", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20286 + }, + { + "path": "322083.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "electronic", + "choice_b": "metal", + "choice_c": "idm", + "choice_d": "house", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20287 + }, + { + "path": "1365376.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "experimental", + "choice_b": "progressive", + "choice_c": "classical", + "choice_d": "ethno", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20288 + }, + { + "path": "1420711.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "swing", + "choice_b": "idm", + "choice_c": "reggae", + "choice_d": "metal", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20289 + }, + { + "path": "1392358.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "atmospheric", + "choice_b": "electropop", + "choice_c": "darkambient", + "choice_d": "reggae", + "answer_gt": "atmospheric", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20290 + }, + { + "path": "967566.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "blues", + "choice_b": "drumnbass", + "choice_c": "classical", + "choice_d": "psychedelic", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20291 + }, + { + "path": "149128.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "classical", + "choice_b": "dance", + "choice_c": "rock", + "choice_d": "minimal", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20292 + }, + { + "path": "771.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "triphop", + "choice_b": "pop", + "choice_c": "hardrock", + "choice_d": "alternative", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20293 + }, + { + "path": "879158.mp3", + "question": "What style of music is this?", + "choice_a": "ambient", + "choice_b": "breakbeat", + "choice_c": "alternative", + "choice_d": "chillout", + "answer_gt": "chillout", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20294 + }, + { + "path": "493296.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "house", + "choice_b": "orchestral", + "choice_c": "industrial", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20295 + }, + { + "path": "6250.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "hiphop", + "choice_b": "soundtrack", + "choice_c": "darkambient", + "choice_d": "trance", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20296 + }, + { + "path": "1058721.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "soundtrack", + "choice_b": "instrumentalpop", + "choice_c": "easylistening", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20297 + }, + { + "path": "988996.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "ambient", + "choice_b": "soundtrack", + "choice_c": "ethno", + "choice_d": "newage", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20298 + }, + { + "path": "1128826.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "chillout", + "choice_b": "punkrock", + "choice_c": "atmospheric", + "choice_d": "country", + "answer_gt": "country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20299 + }, + { + "path": "1329120.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "hardrock", + "choice_b": "chillout", + "choice_c": "rocknroll", + "choice_d": "industrial", + "answer_gt": "rocknroll", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20300 + }, + { + "path": "1084130.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "ambient", + "choice_b": "blues", + "choice_c": "pop", + "choice_d": "orchestral", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20301 + }, + { + "path": "965308.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "bossanova", + "choice_b": "rocknroll", + "choice_c": "classical", + "choice_d": "industrial", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20302 + }, + { + "path": "1085248.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "atmospheric", + "choice_b": "soul", + "choice_c": "classical", + "choice_d": "soundtrack", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20303 + }, + { + "path": "269995.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "electronic", + "choice_b": "classical", + "choice_c": "metal", + "choice_d": "rock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20304 + }, + { + "path": "979865.mp3", + "question": "specify the music genre here?", + "choice_a": "ambient", + "choice_b": "funk", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20305 + }, + { + "path": "1119035.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "acidjazz", + "choice_b": "rap", + "choice_c": "electronic", + "choice_d": "house", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20306 + }, + { + "path": "1374144.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "electronic", + "choice_b": "techno", + "choice_c": "bossanova", + "choice_d": "acidjazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20307 + }, + { + "path": "253879.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "electronic", + "choice_b": "rock", + "choice_c": "progressive", + "choice_d": "world", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20308 + }, + { + "path": "1098745.mp3", + "question": "what's the genre of this music?", + "choice_a": "jazz", + "choice_b": "funk", + "choice_c": "latin", + "choice_d": "hard", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20309 + }, + { + "path": "208278.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "electronic", + "choice_b": "experimental", + "choice_c": "electronica", + "choice_d": "rock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20310 + }, + { + "path": "502325.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "rock", + "choice_b": "classical", + "choice_c": "acidjazz", + "choice_d": "blues", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20311 + }, + { + "path": "1136680.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "popfolk", + "choice_b": "rocknroll", + "choice_c": "rock", + "choice_d": "house", + "answer_gt": "popfolk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20312 + }, + { + "path": "1350826.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "soundtrack", + "choice_b": "trance", + "choice_c": "folk", + "choice_d": "swing", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20313 + }, + { + "path": "1420534.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "darkambient", + "choice_b": "classical", + "choice_c": "drumnbass", + "choice_d": "industrial", + "answer_gt": "drumnbass", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20314 + }, + { + "path": "285623.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "pop", + "choice_b": "experimental", + "choice_c": "ethno", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20315 + }, + { + "path": "1076489.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "rocknroll", + "choice_b": "hardrock", + "choice_c": "easylistening", + "choice_d": "minimal", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20316 + }, + { + "path": "18499.mp3", + "question": "What style of music is this?", + "choice_a": "rocknroll", + "choice_b": "chanson", + "choice_c": "electronic", + "choice_d": "hiphop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20317 + }, + { + "path": "278158.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electronica", + "choice_b": "hiphop", + "choice_c": "classical", + "choice_d": "rap", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20318 + }, + { + "path": "23198.mp3", + "question": "classify the genre of this music?", + "choice_a": "classical", + "choice_b": "trance", + "choice_c": "reggae", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20319 + }, + { + "path": "108751.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "ambient", + "choice_b": "rnb", + "choice_c": "soul", + "choice_d": "rocknroll", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20320 + }, + { + "path": "1350828.mp3", + "question": "detect the genre of this music?", + "choice_a": "soundtrack", + "choice_b": "newwave", + "choice_c": "psychedelic", + "choice_d": "swing", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20321 + }, + { + "path": "295914.mp3", + "question": "what's the genre of this music?", + "choice_a": "newwave", + "choice_b": "atmospheric", + "choice_c": "soundtrack", + "choice_d": "triphop", + "answer_gt": "triphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20322 + }, + { + "path": "434889.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "poprock", + "choice_b": "pop", + "choice_c": "darkambient", + "choice_d": "rock", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20323 + }, + { + "path": "969345.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "chanson", + "choice_b": "easylistening", + "choice_c": "minimal", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20324 + }, + { + "path": "1362269.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "indie", + "choice_b": "electronic", + "choice_c": "orchestral", + "choice_d": "house", + "answer_gt": "orchestral", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20325 + }, + { + "path": "944148.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "hardrock", + "choice_b": "swing", + "choice_c": "classical", + "choice_d": "house", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20326 + }, + { + "path": "147594.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "instrumentalpop", + "choice_b": "country", + "choice_c": "classical", + "choice_d": "rap", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20327 + }, + { + "path": "1105859.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "rocknroll", + "choice_b": "punkrock", + "choice_c": "poprock", + "choice_d": "indie", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20328 + }, + { + "path": "133671.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "dance", + "choice_b": "electropop", + "choice_c": "classical", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20329 + }, + { + "path": "763.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "pop", + "choice_b": "progressive", + "choice_c": "rap", + "choice_d": "atmospheric", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20330 + }, + { + "path": "1350986.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "classical", + "choice_b": "latin", + "choice_c": "newage", + "choice_d": "jazz", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20331 + }, + { + "path": "1272783.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "funk", + "choice_b": "dance", + "choice_c": "alternative", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20332 + }, + { + "path": "1126860.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "funk", + "choice_b": "rock", + "choice_c": "hiphop", + "choice_d": "bossanova", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20333 + }, + { + "path": "1348314.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "rocknroll", + "choice_b": "soundtrack", + "choice_c": "ethno", + "choice_d": "jazz", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20334 + }, + { + "path": "1126846.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "classical", + "choice_b": "acidjazz", + "choice_c": "newage", + "choice_d": "rocknroll", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20335 + }, + { + "path": "1062856.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "world", + "choice_b": "latin", + "choice_c": "poprock", + "choice_d": "minimal", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20336 + }, + { + "path": "1077384.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "swing", + "choice_b": "folk", + "choice_c": "bossanova", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20337 + }, + { + "path": "967565.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "soundtrack", + "choice_b": "drumnbass", + "choice_c": "world", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20338 + }, + { + "path": "224476.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "electronic", + "choice_b": "punkrock", + "choice_c": "blues", + "choice_d": "electronica", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20339 + }, + { + "path": "1154297.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "country", + "choice_b": "triphop", + "choice_c": "ambient", + "choice_d": "indie", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20340 + }, + { + "path": "1349029.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "poprock", + "choice_b": "drumnbass", + "choice_c": "chillout", + "choice_d": "alternative", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20341 + }, + { + "path": "1351055.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "techno", + "choice_b": "soundtrack", + "choice_c": "popfolk", + "choice_d": "folk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20342 + }, + { + "path": "1037363.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "celtic", + "choice_b": "soundtrack", + "choice_c": "ethno", + "choice_d": "drumnbass", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20343 + }, + { + "path": "1116711.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "poprock", + "choice_b": "country", + "choice_c": "breakbeat", + "choice_d": "minimal", + "answer_gt": "country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20344 + }, + { + "path": "431432.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "classical", + "choice_b": "breakbeat", + "choice_c": "minimal", + "choice_d": "orchestral", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20345 + }, + { + "path": "202608.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electronica", + "choice_b": "funk", + "choice_c": "rocknroll", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20346 + }, + { + "path": "245735.mp3", + "question": "What's the classification of this music?", + "choice_a": "atmospheric", + "choice_b": "electronic", + "choice_c": "ethno", + "choice_d": "acidjazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20347 + }, + { + "path": "718561.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "ambient", + "choice_b": "rap", + "choice_c": "punkrock", + "choice_d": "orchestral", + "answer_gt": "rap", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20348 + }, + { + "path": "1050995.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "folk", + "choice_b": "electronic", + "choice_c": "popfolk", + "choice_d": "funk", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20349 + }, + { + "path": "276976.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "instrumentalpop", + "choice_b": "dance", + "choice_c": "rnb", + "choice_d": "hardrock", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20350 + }, + { + "path": "965436.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "classical", + "choice_b": "latin", + "choice_c": "punkrock", + "choice_d": "rock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20351 + }, + { + "path": "173499.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "rock", + "choice_b": "bossanova", + "choice_c": "newage", + "choice_d": "industrial", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20352 + }, + { + "path": "156971.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "hiphop", + "choice_b": "chillout", + "choice_c": "alternative", + "choice_d": "funk", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20353 + }, + { + "path": "1089272.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "rock", + "choice_b": "classical", + "choice_c": "ambient", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20354 + }, + { + "path": "277866.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "easylistening", + "choice_b": "acidjazz", + "choice_c": "poprock", + "choice_d": "reggae", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20355 + }, + { + "path": "221635.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "rock", + "choice_b": "newage", + "choice_c": "funk", + "choice_d": "swing", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20356 + }, + { + "path": "1393671.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "alternative", + "choice_b": "popfolk", + "choice_c": "newwave", + "choice_d": "minimal", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20357 + }, + { + "path": "965316.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "classical", + "choice_b": "folk", + "choice_c": "rnb", + "choice_d": "bossanova", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20358 + }, + { + "path": "211838.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "house", + "choice_b": "triphop", + "choice_c": "celtic", + "choice_d": "psychedelic", + "answer_gt": "triphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20359 + }, + { + "path": "1393670.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "industrial", + "choice_b": "electropop", + "choice_c": "alternative", + "choice_d": "latin", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20360 + }, + { + "path": "815425.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "triphop", + "choice_b": "newage", + "choice_c": "bossanova", + "choice_d": "breakbeat", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20361 + }, + { + "path": "374973.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "reggae", + "choice_b": "bossanova", + "choice_c": "jazz", + "choice_d": "country", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20362 + }, + { + "path": "914774.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "folk", + "choice_b": "bossanova", + "choice_c": "soundtrack", + "choice_d": "hiphop", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20363 + }, + { + "path": "161539.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "jazz", + "choice_b": "celtic", + "choice_c": "chanson", + "choice_d": "acidjazz", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20364 + }, + { + "path": "1127774.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "psychedelic", + "choice_b": "ethno", + "choice_c": "hard", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20365 + }, + { + "path": "1134764.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "drumnbass", + "choice_b": "poprock", + "choice_c": "electronic", + "choice_d": "reggae", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20366 + }, + { + "path": "1350982.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "alternative", + "choice_b": "soundtrack", + "choice_c": "trance", + "choice_d": "progressive", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20367 + }, + { + "path": "149091.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "celtic", + "choice_b": "rock", + "choice_c": "rocknroll", + "choice_d": "bossanova", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20368 + }, + { + "path": "1121562.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "jazz", + "choice_b": "newwave", + "choice_c": "ambient", + "choice_d": "dance", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20369 + }, + { + "path": "1371842.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "celtic", + "choice_b": "latin", + "choice_c": "techno", + "choice_d": "progressive", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20370 + }, + { + "path": "1350141.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "hard", + "choice_b": "pop", + "choice_c": "soundtrack", + "choice_d": "progressive", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20371 + }, + { + "path": "119373.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "orchestral", + "choice_b": "pop", + "choice_c": "funk", + "choice_d": "darkambient", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20372 + }, + { + "path": "208193.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "jazz", + "choice_b": "pop", + "choice_c": "rock", + "choice_d": "experimental", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20373 + }, + { + "path": "1354586.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "soundtrack", + "choice_b": "breakbeat", + "choice_c": "chanson", + "choice_d": "blues", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20374 + }, + { + "path": "1354150.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "industrial", + "choice_b": "hiphop", + "choice_c": "experimental", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20375 + }, + { + "path": "1214584.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "soul", + "choice_b": "indie", + "choice_c": "classical", + "choice_d": "electronica", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20376 + }, + { + "path": "1173147.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "bossanova", + "choice_b": "alternative", + "choice_c": "dance", + "choice_d": "hiphop", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20377 + }, + { + "path": "165331.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "triphop", + "choice_b": "soul", + "choice_c": "electronic", + "choice_d": "newwave", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20378 + }, + { + "path": "253265.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "techno", + "choice_b": "breakbeat", + "choice_c": "rap", + "choice_d": "industrial", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20379 + }, + { + "path": "1062503.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "ambient", + "choice_b": "country", + "choice_c": "bossanova", + "choice_d": "electronica", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20380 + }, + { + "path": "147669.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "latin", + "choice_b": "classical", + "choice_c": "indie", + "choice_d": "blues", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20381 + }, + { + "path": "1400514.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "minimal", + "choice_b": "soul", + "choice_c": "hiphop", + "choice_d": "ethno", + "answer_gt": "minimal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20382 + }, + { + "path": "1043269.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electronic", + "choice_b": "indie", + "choice_c": "pop", + "choice_d": "instrumentalpop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20383 + }, + { + "path": "1350211.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "hardrock", + "choice_b": "breakbeat", + "choice_c": "rnb", + "choice_d": "experimental", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20384 + }, + { + "path": "1354437.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "progressive", + "choice_b": "hardrock", + "choice_c": "soundtrack", + "choice_d": "house", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20385 + }, + { + "path": "1349962.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "soundtrack", + "choice_b": "funk", + "choice_c": "trance", + "choice_d": "folk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20386 + }, + { + "path": "1363401.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "hard", + "choice_b": "funk", + "choice_c": "experimental", + "choice_d": "chanson", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20387 + }, + { + "path": "14016.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "world", + "choice_b": "classical", + "choice_c": "reggae", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20388 + }, + { + "path": "211827.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "alternative", + "choice_b": "electronic", + "choice_c": "chanson", + "choice_d": "popfolk", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20389 + }, + { + "path": "1074444.mp3", + "question": "What's the classification of this music?", + "choice_a": "techno", + "choice_b": "minimal", + "choice_c": "house", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20390 + }, + { + "path": "1332.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "rocknroll", + "choice_b": "electronic", + "choice_c": "instrumentalpop", + "choice_d": "jazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20391 + }, + { + "path": "919997.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "dance", + "choice_b": "rnb", + "choice_c": "newage", + "choice_d": "house", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20392 + }, + { + "path": "757498.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "rock", + "choice_b": "triphop", + "choice_c": "soul", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20393 + }, + { + "path": "1354578.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "newwave", + "choice_b": "soundtrack", + "choice_c": "classical", + "choice_d": "alternative", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20394 + }, + { + "path": "164446.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "darkambient", + "choice_b": "pop", + "choice_c": "metal", + "choice_d": "atmospheric", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20395 + }, + { + "path": "84498.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "reggae", + "choice_b": "experimental", + "choice_c": "psychedelic", + "choice_d": "bossanova", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20396 + }, + { + "path": "22608.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electronic", + "choice_b": "orchestral", + "choice_c": "instrumentalpop", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20397 + }, + { + "path": "1067235.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "idm", + "choice_b": "blues", + "choice_c": "popfolk", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20398 + }, + { + "path": "1344558.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "country", + "choice_b": "darkambient", + "choice_c": "newage", + "choice_d": "techno", + "answer_gt": "darkambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20399 + }, + { + "path": "1347896.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "country", + "choice_b": "classical", + "choice_c": "industrial", + "choice_d": "chillout", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20400 + }, + { + "path": "458679.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "electronic", + "choice_b": "indie", + "choice_c": "soundtrack", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20401 + }, + { + "path": "1133084.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "classical", + "choice_b": "psychedelic", + "choice_c": "rap", + "choice_d": "hiphop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20402 + }, + { + "path": "925530.mp3", + "question": "specify the music genre here?", + "choice_a": "classical", + "choice_b": "darkambient", + "choice_c": "hiphop", + "choice_d": "hard", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20403 + }, + { + "path": "395258.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "pop", + "choice_b": "indie", + "choice_c": "breakbeat", + "choice_d": "techno", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20404 + }, + { + "path": "1349988.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "experimental", + "choice_b": "classical", + "choice_c": "soul", + "choice_d": "funk", + "answer_gt": "experimental", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20405 + }, + { + "path": "381706.mp3", + "question": "classify the genre of this music?", + "choice_a": "reggae", + "choice_b": "folk", + "choice_c": "trance", + "choice_d": "rock", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20406 + }, + { + "path": "1002223.mp3", + "question": "classify the genre of this music?", + "choice_a": "jazz", + "choice_b": "alternative", + "choice_c": "world", + "choice_d": "minimal", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20407 + }, + { + "path": "972774.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "industrial", + "choice_b": "electronic", + "choice_c": "drumnbass", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20408 + }, + { + "path": "184207.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "rocknroll", + "choice_b": "classical", + "choice_c": "celtic", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20409 + }, + { + "path": "1348859.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "soul", + "choice_b": "rnb", + "choice_c": "poprock", + "choice_d": "breakbeat", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20410 + }, + { + "path": "1348855.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "metal", + "choice_b": "rnb", + "choice_c": "rap", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20411 + }, + { + "path": "969347.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "classical", + "choice_b": "psychedelic", + "choice_c": "blues", + "choice_d": "trance", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20412 + }, + { + "path": "1348927.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "reggae", + "choice_b": "poprock", + "choice_c": "house", + "choice_d": "minimal", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20413 + }, + { + "path": "769.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "ambient", + "choice_b": "breakbeat", + "choice_c": "pop", + "choice_d": "alternative", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20414 + }, + { + "path": "1267263.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "poprock", + "choice_b": "techno", + "choice_c": "hiphop", + "choice_d": "trance", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20415 + }, + { + "path": "1227499.mp3", + "question": "classify the genre of this music?", + "choice_a": "soundtrack", + "choice_b": "dance", + "choice_c": "trance", + "choice_d": "metal", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20416 + }, + { + "path": "14021.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "classical", + "choice_b": "instrumentalpop", + "choice_c": "rock", + "choice_d": "trance", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20417 + }, + { + "path": "965760.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "orchestral", + "choice_b": "psychedelic", + "choice_c": "latin", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20418 + }, + { + "path": "1026823.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "classical", + "choice_b": "trance", + "choice_c": "newwave", + "choice_d": "punkrock", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20419 + }, + { + "path": "575216.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "popfolk", + "choice_b": "trance", + "choice_c": "bossanova", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20420 + }, + { + "path": "925526.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "classical", + "choice_b": "swing", + "choice_c": "soul", + "choice_d": "rocknroll", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20421 + }, + { + "path": "757496.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "country", + "choice_b": "classical", + "choice_c": "rocknroll", + "choice_d": "reggae", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20422 + }, + { + "path": "363085.mp3", + "question": "What's the classification of this music?", + "choice_a": "minimal", + "choice_b": "darkambient", + "choice_c": "electronic", + "choice_d": "jazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20423 + }, + { + "path": "944153.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "classical", + "choice_b": "hiphop", + "choice_c": "latin", + "choice_d": "darkambient", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20424 + }, + { + "path": "86742.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "easylistening", + "choice_b": "folk", + "choice_c": "newage", + "choice_d": "alternative", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20425 + }, + { + "path": "216409.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "instrumentalpop", + "choice_b": "funk", + "choice_c": "rock", + "choice_d": "latin", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20426 + }, + { + "path": "1037344.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "folk", + "choice_b": "celtic", + "choice_c": "bossanova", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20427 + }, + { + "path": "1286511.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "instrumentalpop", + "choice_b": "chanson", + "choice_c": "world", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20428 + }, + { + "path": "1127777.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "rnb", + "choice_b": "folk", + "choice_c": "instrumentalpop", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20429 + }, + { + "path": "401458.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "electronic", + "choice_b": "techno", + "choice_c": "idm", + "choice_d": "trance", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20430 + }, + { + "path": "1400474.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "country", + "choice_b": "alternative", + "choice_c": "electropop", + "choice_d": "reggae", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20431 + }, + { + "path": "328650.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "blues", + "choice_b": "hiphop", + "choice_c": "house", + "choice_d": "rocknroll", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20432 + }, + { + "path": "257599.mp3", + "question": "specify the music genre here?", + "choice_a": "electropop", + "choice_b": "alternative", + "choice_c": "easylistening", + "choice_d": "hard", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20433 + }, + { + "path": "1408401.mp3", + "question": "what's the genre of this music?", + "choice_a": "acidjazz", + "choice_b": "rnb", + "choice_c": "celtic", + "choice_d": "jazz", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20434 + }, + { + "path": "1121130.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "punkrock", + "choice_b": "alternative", + "choice_c": "pop", + "choice_d": "rocknroll", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20435 + }, + { + "path": "1108515.mp3", + "question": "specify the music genre here?", + "choice_a": "rnb", + "choice_b": "soul", + "choice_c": "psychedelic", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20436 + }, + { + "path": "508155.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "dance", + "choice_b": "hardrock", + "choice_c": "rock", + "choice_d": "breakbeat", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20437 + }, + { + "path": "324378.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "soul", + "choice_b": "industrial", + "choice_c": "easylistening", + "choice_d": "instrumentalpop", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20438 + }, + { + "path": "1350135.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "swing", + "choice_b": "newage", + "choice_c": "soundtrack", + "choice_d": "punkrock", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20439 + }, + { + "path": "191405.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "industrial", + "choice_b": "newage", + "choice_c": "punkrock", + "choice_d": "acidjazz", + "answer_gt": "industrial", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20440 + }, + { + "path": "210083.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "electropop", + "choice_b": "triphop", + "choice_c": "ethno", + "choice_d": "industrial", + "answer_gt": "triphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20441 + }, + { + "path": "1026820.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "celtic", + "choice_b": "trance", + "choice_c": "swing", + "choice_d": "country", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20442 + }, + { + "path": "842758.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "soundtrack", + "choice_b": "jazz", + "choice_c": "classical", + "choice_d": "dance", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20443 + }, + { + "path": "83425.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "rnb", + "choice_b": "industrial", + "choice_c": "electronic", + "choice_d": "atmospheric", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20444 + }, + { + "path": "242572.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "classical", + "choice_b": "psychedelic", + "choice_c": "orchestral", + "choice_d": "easylistening", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20445 + }, + { + "path": "13284.mp3", + "question": "what's the genre of this music?", + "choice_a": "rock", + "choice_b": "punkrock", + "choice_c": "chillout", + "choice_d": "atmospheric", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20446 + }, + { + "path": "226684.mp3", + "question": "What's the classification of this music?", + "choice_a": "hardrock", + "choice_b": "electronica", + "choice_c": "rock", + "choice_d": "psychedelic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20447 + }, + { + "path": "965315.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "newwave", + "choice_b": "atmospheric", + "choice_c": "classical", + "choice_d": "metal", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20448 + }, + { + "path": "985179.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "hard", + "choice_b": "electronica", + "choice_c": "electronic", + "choice_d": "acidjazz", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20449 + }, + { + "path": "1398526.mp3", + "question": "classify the genre of this music?", + "choice_a": "pop", + "choice_b": "dance", + "choice_c": "ambient", + "choice_d": "country", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20450 + }, + { + "path": "83423.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "punkrock", + "choice_b": "electronic", + "choice_c": "soundtrack", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20451 + }, + { + "path": "208294.mp3", + "question": "identify the music genre of this track.", + "choice_a": "rock", + "choice_b": "pop", + "choice_c": "punkrock", + "choice_d": "soul", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20452 + }, + { + "path": "42320.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "dance", + "choice_b": "reggae", + "choice_c": "psychedelic", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20453 + }, + { + "path": "1089266.mp3", + "question": "identify the music genre of this track.", + "choice_a": "blues", + "choice_b": "latin", + "choice_c": "progressive", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20454 + }, + { + "path": "1062502.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "electronica", + "choice_b": "ambient", + "choice_c": "breakbeat", + "choice_d": "metal", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20455 + }, + { + "path": "127600.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "folk", + "choice_b": "psychedelic", + "choice_c": "darkambient", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20456 + }, + { + "path": "873137.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "atmospheric", + "choice_b": "classical", + "choice_c": "psychedelic", + "choice_d": "electropop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20457 + }, + { + "path": "1076572.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "breakbeat", + "choice_b": "industrial", + "choice_c": "popfolk", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20458 + }, + { + "path": "1353448.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "orchestral", + "choice_b": "blues", + "choice_c": "soundtrack", + "choice_d": "rnb", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20459 + }, + { + "path": "1062500.mp3", + "question": "detect the genre of this music?", + "choice_a": "ambient", + "choice_b": "indie", + "choice_c": "celtic", + "choice_d": "minimal", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20460 + }, + { + "path": "767.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "celtic", + "choice_b": "newwave", + "choice_c": "breakbeat", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20461 + }, + { + "path": "965318.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "electronica", + "choice_b": "hiphop", + "choice_c": "celtic", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20462 + }, + { + "path": "110666.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "acidjazz", + "choice_b": "progressive", + "choice_c": "soundtrack", + "choice_d": "trance", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20463 + }, + { + "path": "149069.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "popfolk", + "choice_b": "rock", + "choice_c": "punkrock", + "choice_d": "atmospheric", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20464 + }, + { + "path": "985173.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "latin", + "choice_b": "minimal", + "choice_c": "electronica", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20465 + }, + { + "path": "1399653.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "world", + "choice_b": "idm", + "choice_c": "triphop", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20466 + }, + { + "path": "1076563.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "industrial", + "choice_b": "world", + "choice_c": "hiphop", + "choice_d": "rnb", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20467 + }, + { + "path": "1126028.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "indie", + "choice_b": "experimental", + "choice_c": "hiphop", + "choice_d": "soul", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20468 + }, + { + "path": "1131682.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "pop", + "choice_b": "instrumentalpop", + "choice_c": "ethno", + "choice_d": "industrial", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20469 + }, + { + "path": "991647.mp3", + "question": "What style of music is this?", + "choice_a": "ambient", + "choice_b": "world", + "choice_c": "newwave", + "choice_d": "idm", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20470 + }, + { + "path": "1348853.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "rnb", + "choice_b": "world", + "choice_c": "poprock", + "choice_d": "electropop", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20471 + }, + { + "path": "240476.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "industrial", + "choice_b": "pop", + "choice_c": "rock", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20472 + }, + { + "path": "458682.mp3", + "question": "specify the music genre here?", + "choice_a": "chanson", + "choice_b": "electronic", + "choice_c": "latin", + "choice_d": "easylistening", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20473 + }, + { + "path": "383976.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "reggae", + "choice_b": "pop", + "choice_c": "hardrock", + "choice_d": "drumnbass", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20474 + }, + { + "path": "1113615.mp3", + "question": "classify the genre of this music?", + "choice_a": "alternative", + "choice_b": "reggae", + "choice_c": "newwave", + "choice_d": "folk", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20475 + }, + { + "path": "1395666.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "pop", + "choice_b": "orchestral", + "choice_c": "jazz", + "choice_d": "hard", + "answer_gt": "hard", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20476 + }, + { + "path": "1383743.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "classical", + "choice_b": "pop", + "choice_c": "hard", + "choice_d": "dance", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20477 + }, + { + "path": "1134327.mp3", + "question": "detect the genre of this music?", + "choice_a": "drumnbass", + "choice_b": "metal", + "choice_c": "pop", + "choice_d": "classical", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20478 + }, + { + "path": "1094524.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "electronic", + "choice_b": "chanson", + "choice_c": "pop", + "choice_d": "funk", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20479 + }, + { + "path": "1035892.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "pop", + "choice_b": "minimal", + "choice_c": "rocknroll", + "choice_d": "indie", + "answer_gt": "indie", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20480 + }, + { + "path": "1127767.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "reggae", + "choice_b": "hiphop", + "choice_c": "minimal", + "choice_d": "atmospheric", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20481 + }, + { + "path": "5339.mp3", + "question": "classify the genre of this music?", + "choice_a": "electronic", + "choice_b": "rnb", + "choice_c": "chanson", + "choice_d": "triphop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20482 + }, + { + "path": "1098858.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "rap", + "choice_b": "ambient", + "choice_c": "hard", + "choice_d": "electronica", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20483 + }, + { + "path": "1226100.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "acidjazz", + "choice_b": "electronic", + "choice_c": "darkambient", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20484 + }, + { + "path": "1406017.mp3", + "question": "specify the music genre here?", + "choice_a": "latin", + "choice_b": "ambient", + "choice_c": "popfolk", + "choice_d": "indie", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20485 + }, + { + "path": "467850.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "chillout", + "choice_b": "hiphop", + "choice_c": "country", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20486 + }, + { + "path": "1072591.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "poprock", + "choice_b": "acidjazz", + "choice_c": "psychedelic", + "choice_d": "idm", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20487 + }, + { + "path": "1107947.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "rock", + "choice_b": "blues", + "choice_c": "electronic", + "choice_d": "experimental", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20488 + }, + { + "path": "42321.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "easylistening", + "choice_b": "country", + "choice_c": "rock", + "choice_d": "pop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20489 + }, + { + "path": "549231.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "pop", + "choice_b": "ethno", + "choice_c": "hiphop", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20490 + }, + { + "path": "847334.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "metal", + "choice_b": "electronic", + "choice_c": "electropop", + "choice_d": "chanson", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20491 + }, + { + "path": "1154516.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "pop", + "choice_b": "funk", + "choice_c": "soundtrack", + "choice_d": "chanson", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20492 + }, + { + "path": "718564.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "reggae", + "choice_b": "triphop", + "choice_c": "indie", + "choice_d": "trance", + "answer_gt": "triphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20493 + }, + { + "path": "1352150.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "ethno", + "choice_b": "soul", + "choice_c": "soundtrack", + "choice_d": "swing", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20494 + }, + { + "path": "161537.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "newwave", + "choice_b": "jazz", + "choice_c": "house", + "choice_d": "triphop", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20495 + }, + { + "path": "1116268.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "electropop", + "choice_b": "drumnbass", + "choice_c": "celtic", + "choice_d": "latin", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20496 + }, + { + "path": "979857.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "bossanova", + "choice_b": "ambient", + "choice_c": "easylistening", + "choice_d": "poprock", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20497 + }, + { + "path": "9600.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "classical", + "choice_b": "alternative", + "choice_c": "pop", + "choice_d": "acidjazz", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20498 + }, + { + "path": "208190.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "breakbeat", + "choice_b": "rock", + "choice_c": "hardrock", + "choice_d": "instrumentalpop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20499 + }, + { + "path": "340491.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "rap", + "choice_b": "easylistening", + "choice_c": "darkambient", + "choice_d": "trance", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20500 + }, + { + "path": "431434.mp3", + "question": "detect the genre of this music?", + "choice_a": "hard", + "choice_b": "experimental", + "choice_c": "funk", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20501 + }, + { + "path": "1403976.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "rock", + "choice_b": "instrumentalpop", + "choice_c": "indie", + "choice_d": "chillout", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20502 + }, + { + "path": "1350307.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "soundtrack", + "choice_b": "popfolk", + "choice_c": "experimental", + "choice_d": "funk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20503 + }, + { + "path": "1286513.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "rap", + "choice_b": "alternative", + "choice_c": "blues", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20504 + }, + { + "path": "1286515.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "rock", + "choice_b": "newage", + "choice_c": "popfolk", + "choice_d": "punkrock", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20505 + }, + { + "path": "881283.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "minimal", + "choice_b": "funk", + "choice_c": "rocknroll", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20506 + }, + { + "path": "1352899.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "newwave", + "choice_b": "electropop", + "choice_c": "acidjazz", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20507 + }, + { + "path": "31511.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "newage", + "choice_b": "house", + "choice_c": "pop", + "choice_d": "hardrock", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20508 + }, + { + "path": "84494.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "hard", + "choice_b": "jazz", + "choice_c": "electronica", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20509 + }, + { + "path": "508153.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "rock", + "choice_b": "psychedelic", + "choice_c": "hardrock", + "choice_d": "newage", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20510 + }, + { + "path": "1290919.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "indie", + "choice_b": "popfolk", + "choice_c": "ethno", + "choice_d": "punkrock", + "answer_gt": "popfolk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20511 + }, + { + "path": "1119919.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "jazz", + "choice_b": "soundtrack", + "choice_c": "country", + "choice_d": "alternative", + "answer_gt": "country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20512 + }, + { + "path": "965314.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "classical", + "choice_b": "easylistening", + "choice_c": "hiphop", + "choice_d": "instrumentalpop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20513 + }, + { + "path": "1182014.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "hardrock", + "choice_b": "electronic", + "choice_c": "hiphop", + "choice_d": "blues", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20514 + }, + { + "path": "431447.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "rocknroll", + "choice_b": "classical", + "choice_c": "newwave", + "choice_d": "alternative", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20515 + }, + { + "path": "128301.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "house", + "choice_b": "drumnbass", + "choice_c": "easylistening", + "choice_d": "orchestral", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20516 + }, + { + "path": "434886.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "pop", + "choice_b": "hard", + "choice_c": "reggae", + "choice_d": "industrial", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20517 + }, + { + "path": "989718.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "electronica", + "choice_b": "classical", + "choice_c": "indie", + "choice_d": "newage", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20518 + }, + { + "path": "1153922.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "darkambient", + "choice_b": "techno", + "choice_c": "psychedelic", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20519 + }, + { + "path": "253755.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "classical", + "choice_b": "electronic", + "choice_c": "rnb", + "choice_d": "rock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20520 + }, + { + "path": "216410.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "electronica", + "choice_b": "electropop", + "choice_c": "chanson", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20521 + }, + { + "path": "1031109.mp3", + "question": "identify the music genre of this track.", + "choice_a": "triphop", + "choice_b": "easylistening", + "choice_c": "electronic", + "choice_d": "rnb", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20522 + }, + { + "path": "208191.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "reggae", + "choice_b": "rock", + "choice_c": "drumnbass", + "choice_d": "latin", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20523 + }, + { + "path": "339430.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "chanson", + "choice_b": "experimental", + "choice_c": "metal", + "choice_d": "rock", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20524 + }, + { + "path": "985175.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "dance", + "choice_b": "idm", + "choice_c": "orchestral", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20525 + }, + { + "path": "762.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "pop", + "choice_b": "chanson", + "choice_c": "atmospheric", + "choice_d": "rocknroll", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20526 + }, + { + "path": "29942.mp3", + "question": "What's the classification of this music?", + "choice_a": "alternative", + "choice_b": "electronic", + "choice_c": "easylistening", + "choice_d": "hardrock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20527 + }, + { + "path": "1330564.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "world", + "choice_b": "soul", + "choice_c": "folk", + "choice_d": "latin", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20528 + }, + { + "path": "1286485.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "hiphop", + "choice_b": "indie", + "choice_c": "newage", + "choice_d": "hardrock", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20529 + }, + { + "path": "734332.mp3", + "question": "what's the genre of this music?", + "choice_a": "psychedelic", + "choice_b": "latin", + "choice_c": "rock", + "choice_d": "jazz", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20530 + }, + { + "path": "1350957.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "soundtrack", + "choice_b": "orchestral", + "choice_c": "reggae", + "choice_d": "electronica", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20531 + }, + { + "path": "40830.mp3", + "question": "specify the music genre here?", + "choice_a": "orchestral", + "choice_b": "idm", + "choice_c": "ethno", + "choice_d": "psychedelic", + "answer_gt": "ethno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20532 + }, + { + "path": "1088579.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "pop", + "choice_b": "punkrock", + "choice_c": "rnb", + "choice_d": "funk", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20533 + }, + { + "path": "1349402.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "swing", + "choice_b": "soundtrack", + "choice_c": "classical", + "choice_d": "chillout", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20534 + }, + { + "path": "1362268.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "poprock", + "choice_b": "soundtrack", + "choice_c": "rocknroll", + "choice_d": "experimental", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20535 + }, + { + "path": "393249.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "classical", + "choice_b": "ambient", + "choice_c": "drumnbass", + "choice_d": "psychedelic", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20536 + }, + { + "path": "1037362.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "popfolk", + "choice_b": "soundtrack", + "choice_c": "atmospheric", + "choice_d": "minimal", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20537 + }, + { + "path": "1173332.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "alternative", + "choice_b": "experimental", + "choice_c": "world", + "choice_d": "dance", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20538 + }, + { + "path": "1093496.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "house", + "choice_b": "triphop", + "choice_c": "trance", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20539 + }, + { + "path": "1257403.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "popfolk", + "choice_b": "chillout", + "choice_c": "newwave", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20540 + }, + { + "path": "119379.mp3", + "question": "What's the classification of this music?", + "choice_a": "pop", + "choice_b": "idm", + "choice_c": "soul", + "choice_d": "latin", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20541 + }, + { + "path": "14013.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "breakbeat", + "choice_b": "darkambient", + "choice_c": "ambient", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20542 + }, + { + "path": "1330559.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "chillout", + "choice_b": "rnb", + "choice_c": "world", + "choice_d": "electronica", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20543 + }, + { + "path": "81080.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "swing", + "choice_b": "progressive", + "choice_c": "drumnbass", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20544 + }, + { + "path": "295780.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "experimental", + "choice_b": "funk", + "choice_c": "soundtrack", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20545 + }, + { + "path": "847335.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "swing", + "choice_b": "pop", + "choice_c": "world", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20546 + }, + { + "path": "1349955.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "soundtrack", + "choice_b": "electronica", + "choice_c": "chillout", + "choice_d": "house", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20547 + }, + { + "path": "1085242.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "classical", + "choice_b": "punkrock", + "choice_c": "electronic", + "choice_d": "rnb", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20548 + }, + { + "path": "1112064.mp3", + "question": "identify the music genre of this track.", + "choice_a": "poprock", + "choice_b": "world", + "choice_c": "alternative", + "choice_d": "funk", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20549 + }, + { + "path": "862191.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "soundtrack", + "choice_b": "drumnbass", + "choice_c": "metal", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20550 + }, + { + "path": "147664.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "world", + "choice_b": "dance", + "choice_c": "rock", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20551 + }, + { + "path": "401453.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "world", + "choice_b": "electronic", + "choice_c": "industrial", + "choice_d": "soul", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20552 + }, + { + "path": "191375.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "industrial", + "choice_b": "pop", + "choice_c": "soul", + "choice_d": "ethno", + "answer_gt": "industrial", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20553 + }, + { + "path": "594846.mp3", + "question": "identify the music genre of this track.", + "choice_a": "soundtrack", + "choice_b": "pop", + "choice_c": "easylistening", + "choice_d": "rnb", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20554 + }, + { + "path": "127559.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "folk", + "choice_b": "rocknroll", + "choice_c": "easylistening", + "choice_d": "celtic", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20555 + }, + { + "path": "965752.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "alternative", + "choice_b": "psychedelic", + "choice_c": "pop", + "choice_d": "electropop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20556 + }, + { + "path": "965758.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "ambient", + "choice_b": "orchestral", + "choice_c": "rock", + "choice_d": "chanson", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20557 + }, + { + "path": "179675.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "jazz", + "choice_b": "atmospheric", + "choice_c": "industrial", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20558 + }, + { + "path": "368179.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "rnb", + "choice_b": "chanson", + "choice_c": "darkambient", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20559 + }, + { + "path": "141085.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "hardrock", + "choice_b": "bossanova", + "choice_c": "newage", + "choice_d": "hiphop", + "answer_gt": "hardrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20560 + }, + { + "path": "1051666.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "dance", + "choice_b": "rock", + "choice_c": "electropop", + "choice_d": "idm", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20561 + }, + { + "path": "1400510.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "industrial", + "choice_b": "hard", + "choice_c": "popfolk", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20562 + }, + { + "path": "1126847.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "psychedelic", + "choice_b": "rap", + "choice_c": "classical", + "choice_d": "swing", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20563 + }, + { + "path": "245758.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "breakbeat", + "choice_b": "atmospheric", + "choice_c": "alternative", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20564 + }, + { + "path": "1270332.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "industrial", + "choice_b": "house", + "choice_c": "reggae", + "choice_d": "orchestral", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20565 + }, + { + "path": "1084954.mp3", + "question": "what's the genre of this music?", + "choice_a": "world", + "choice_b": "hardrock", + "choice_c": "psychedelic", + "choice_d": "newwave", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20566 + }, + { + "path": "285674.mp3", + "question": "detect the genre of this music?", + "choice_a": "industrial", + "choice_b": "pop", + "choice_c": "classical", + "choice_d": "idm", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20567 + }, + { + "path": "1061177.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "classical", + "choice_b": "punkrock", + "choice_c": "chillout", + "choice_d": "newwave", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20568 + }, + { + "path": "1350849.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "world", + "choice_b": "experimental", + "choice_c": "swing", + "choice_d": "minimal", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20569 + }, + { + "path": "1096078.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "instrumentalpop", + "choice_b": "drumnbass", + "choice_c": "minimal", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20570 + }, + { + "path": "76951.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "newage", + "choice_b": "soundtrack", + "choice_c": "bossanova", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20571 + }, + { + "path": "1128817.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "ethno", + "choice_b": "funk", + "choice_c": "orchestral", + "choice_d": "latin", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20572 + }, + { + "path": "984997.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "soul", + "choice_b": "newage", + "choice_c": "house", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20573 + }, + { + "path": "1128819.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "latin", + "choice_b": "rocknroll", + "choice_c": "country", + "choice_d": "rap", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20574 + }, + { + "path": "1163971.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "folk", + "choice_b": "minimal", + "choice_c": "experimental", + "choice_d": "electronica", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20575 + }, + { + "path": "532550.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "reggae", + "choice_b": "house", + "choice_c": "electronic", + "choice_d": "chanson", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20576 + }, + { + "path": "949647.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "classical", + "choice_b": "popfolk", + "choice_c": "metal", + "choice_d": "hiphop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20577 + }, + { + "path": "106894.mp3", + "question": "what's the genre of this music?", + "choice_a": "indie", + "choice_b": "rnb", + "choice_c": "electronic", + "choice_d": "world", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20578 + }, + { + "path": "212011.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "hard", + "choice_b": "techno", + "choice_c": "acidjazz", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20579 + }, + { + "path": "1126027.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "breakbeat", + "choice_b": "house", + "choice_c": "funk", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20580 + }, + { + "path": "1227495.mp3", + "question": "specify the music genre here?", + "choice_a": "electronica", + "choice_b": "poprock", + "choice_c": "classical", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20581 + }, + { + "path": "1086429.mp3", + "question": "what's the genre of this music?", + "choice_a": "newwave", + "choice_b": "jazz", + "choice_c": "ambient", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20582 + }, + { + "path": "1197945.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "metal", + "choice_b": "electronic", + "choice_c": "electropop", + "choice_d": "punkrock", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20583 + }, + { + "path": "163769.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "industrial", + "choice_b": "soul", + "choice_c": "house", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20584 + }, + { + "path": "879155.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "chillout", + "choice_b": "celtic", + "choice_c": "rap", + "choice_d": "industrial", + "answer_gt": "chillout", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20585 + }, + { + "path": "264302.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "electronic", + "choice_b": "pop", + "choice_c": "hiphop", + "choice_d": "techno", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20586 + }, + { + "path": "434858.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "soundtrack", + "choice_b": "instrumentalpop", + "choice_c": "drumnbass", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20587 + }, + { + "path": "1086691.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "ambient", + "choice_b": "pop", + "choice_c": "classical", + "choice_d": "minimal", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20588 + }, + { + "path": "1257397.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "poprock", + "choice_b": "folk", + "choice_c": "bossanova", + "choice_d": "punkrock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20589 + }, + { + "path": "295920.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "triphop", + "choice_b": "chanson", + "choice_c": "ethno", + "choice_d": "idm", + "answer_gt": "triphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20590 + }, + { + "path": "82882.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "funk", + "choice_b": "electronic", + "choice_c": "reggae", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20591 + }, + { + "path": "1400517.mp3", + "question": "What style of music is this?", + "choice_a": "swing", + "choice_b": "electronic", + "choice_c": "easylistening", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20592 + }, + { + "path": "1353542.mp3", + "question": "classify the genre of this music?", + "choice_a": "breakbeat", + "choice_b": "soundtrack", + "choice_c": "orchestral", + "choice_d": "rnb", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20593 + }, + { + "path": "130316.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "ambient", + "choice_b": "hard", + "choice_c": "celtic", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20594 + }, + { + "path": "208188.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "classical", + "choice_b": "rock", + "choice_c": "dance", + "choice_d": "reggae", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20595 + }, + { + "path": "196320.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "popfolk", + "choice_b": "chillout", + "choice_c": "hiphop", + "choice_d": "psychedelic", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20596 + }, + { + "path": "88873.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "electronic", + "choice_b": "pop", + "choice_c": "indie", + "choice_d": "funk", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20597 + }, + { + "path": "431344.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "pop", + "choice_b": "classical", + "choice_c": "ambient", + "choice_d": "instrumentalpop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20598 + }, + { + "path": "395269.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "electronica", + "choice_b": "triphop", + "choice_c": "pop", + "choice_d": "soul", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20599 + }, + { + "path": "1119431.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "blues", + "choice_b": "soul", + "choice_c": "rap", + "choice_d": "industrial", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20600 + }, + { + "path": "1257399.mp3", + "question": "specify the music genre here?", + "choice_a": "popfolk", + "choice_b": "poprock", + "choice_c": "dance", + "choice_d": "soundtrack", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20601 + }, + { + "path": "1167071.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "country", + "choice_b": "funk", + "choice_c": "jazz", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20602 + }, + { + "path": "401455.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "acidjazz", + "choice_b": "chanson", + "choice_c": "electronic", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20603 + }, + { + "path": "224477.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "folk", + "choice_b": "electronic", + "choice_c": "trance", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20604 + }, + { + "path": "1418299.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "atmospheric", + "choice_b": "popfolk", + "choice_c": "progressive", + "choice_d": "idm", + "answer_gt": "atmospheric", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20605 + }, + { + "path": "1098746.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "reggae", + "choice_b": "electronic", + "choice_c": "celtic", + "choice_d": "chanson", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20606 + }, + { + "path": "32785.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "classical", + "choice_b": "folk", + "choice_c": "soul", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20607 + }, + { + "path": "1127771.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "funk", + "choice_b": "rap", + "choice_c": "hiphop", + "choice_d": "easylistening", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20608 + }, + { + "path": "81921.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "blues", + "choice_b": "classical", + "choice_c": "pop", + "choice_d": "jazz", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20609 + }, + { + "path": "1344546.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "jazz", + "choice_b": "darkambient", + "choice_c": "chillout", + "choice_d": "swing", + "answer_gt": "darkambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20610 + }, + { + "path": "285665.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "blues", + "choice_b": "acidjazz", + "choice_c": "idm", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20611 + }, + { + "path": "1026816.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "trance", + "choice_b": "world", + "choice_c": "instrumentalpop", + "choice_d": "house", + "answer_gt": "trance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20612 + }, + { + "path": "208184.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "newwave", + "choice_b": "rock", + "choice_c": "orchestral", + "choice_d": "rocknroll", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20613 + }, + { + "path": "317054.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "electronic", + "choice_b": "country", + "choice_c": "rocknroll", + "choice_d": "soul", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20614 + }, + { + "path": "1007831.mp3", + "question": "What style of music is this?", + "choice_a": "poprock", + "choice_b": "breakbeat", + "choice_c": "progressive", + "choice_d": "classical", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20615 + }, + { + "path": "1344545.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "darkambient", + "choice_b": "country", + "choice_c": "indie", + "choice_d": "funk", + "answer_gt": "funk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20616 + }, + { + "path": "362492.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "electropop", + "choice_b": "pop", + "choice_c": "progressive", + "choice_d": "breakbeat", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20617 + }, + { + "path": "979867.mp3", + "question": "detect the genre of this music?", + "choice_a": "popfolk", + "choice_b": "newage", + "choice_c": "indie", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20618 + }, + { + "path": "5341.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "dance", + "choice_b": "ambient", + "choice_c": "jazz", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20619 + }, + { + "path": "1127768.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "blues", + "choice_b": "minimal", + "choice_c": "hiphop", + "choice_d": "chanson", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20620 + }, + { + "path": "799917.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "punkrock", + "choice_b": "easylistening", + "choice_c": "bossanova", + "choice_d": "darkambient", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20621 + }, + { + "path": "1086682.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "rnb", + "choice_b": "classical", + "choice_c": "electronic", + "choice_d": "latin", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20622 + }, + { + "path": "6698.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "bossanova", + "choice_b": "electronic", + "choice_c": "psychedelic", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20623 + }, + { + "path": "81915.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "hardrock", + "choice_b": "trance", + "choice_c": "pop", + "choice_d": "bossanova", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20624 + }, + { + "path": "218524.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "ambient", + "choice_b": "alternative", + "choice_c": "orchestral", + "choice_d": "industrial", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20625 + }, + { + "path": "1272492.mp3", + "question": "what's the genre of this music?", + "choice_a": "progressive", + "choice_b": "ambient", + "choice_c": "ethno", + "choice_d": "classical", + "answer_gt": "progressive", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20626 + }, + { + "path": "1086431.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "popfolk", + "choice_b": "hardrock", + "choice_c": "hiphop", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20627 + }, + { + "path": "1351058.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "bossanova", + "choice_b": "blues", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20628 + }, + { + "path": "1156579.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "darkambient", + "choice_b": "breakbeat", + "choice_c": "soundtrack", + "choice_d": "country", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20629 + }, + { + "path": "474694.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "metal", + "choice_b": "acidjazz", + "choice_c": "hard", + "choice_d": "hiphop", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20630 + }, + { + "path": "1267670.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "classical", + "choice_b": "latin", + "choice_c": "drumnbass", + "choice_d": "darkambient", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20631 + }, + { + "path": "1415804.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "hiphop", + "choice_b": "alternative", + "choice_c": "electropop", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20632 + }, + { + "path": "1163965.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "rock", + "choice_b": "ambient", + "choice_c": "folk", + "choice_d": "breakbeat", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20633 + }, + { + "path": "242568.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "latin", + "choice_b": "techno", + "choice_c": "classical", + "choice_d": "punkrock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20634 + }, + { + "path": "1372924.mp3", + "question": "what's the genre of this music?", + "choice_a": "rock", + "choice_b": "instrumentalpop", + "choice_c": "celtic", + "choice_d": "hiphop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20635 + }, + { + "path": "221631.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "house", + "choice_b": "electronica", + "choice_c": "techno", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20636 + }, + { + "path": "1083529.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "rap", + "choice_b": "house", + "choice_c": "pop", + "choice_d": "progressive", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20637 + }, + { + "path": "772.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "pop", + "choice_b": "latin", + "choice_c": "hiphop", + "choice_d": "newwave", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20638 + }, + { + "path": "224499.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "jazz", + "choice_b": "progressive", + "choice_c": "electronic", + "choice_d": "indie", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20639 + }, + { + "path": "161546.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "jazz", + "choice_b": "soul", + "choice_c": "breakbeat", + "choice_d": "hard", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20640 + }, + { + "path": "1076580.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "easylistening", + "choice_b": "hiphop", + "choice_c": "dance", + "choice_d": "reggae", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20641 + }, + { + "path": "948722.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "psychedelic", + "choice_b": "folk", + "choice_c": "classical", + "choice_d": "world", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20642 + }, + { + "path": "251683.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "instrumentalpop", + "choice_b": "world", + "choice_c": "idm", + "choice_d": "techno", + "answer_gt": "instrumentalpop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20643 + }, + { + "path": "237606.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "trance", + "choice_b": "swing", + "choice_c": "latin", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20644 + }, + { + "path": "127687.mp3", + "question": "What's the musical style represented in this clip?", + "choice_a": "electropop", + "choice_b": "easylistening", + "choice_c": "soundtrack", + "choice_d": "triphop", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20645 + }, + { + "path": "1257396.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "poprock", + "choice_b": "blues", + "choice_c": "rock", + "choice_d": "newwave", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20646 + }, + { + "path": "949646.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "classical", + "choice_b": "jazz", + "choice_c": "hard", + "choice_d": "electronica", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20647 + }, + { + "path": "1349531.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "orchestral", + "choice_b": "drumnbass", + "choice_c": "chillout", + "choice_d": "reggae", + "answer_gt": "orchestral", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20648 + }, + { + "path": "976711.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "classical", + "choice_b": "easylistening", + "choice_c": "idm", + "choice_d": "drumnbass", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20649 + }, + { + "path": "979837.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "swing", + "choice_b": "techno", + "choice_c": "darkambient", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20650 + }, + { + "path": "253236.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "swing", + "choice_b": "bossanova", + "choice_c": "hard", + "choice_d": "techno", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20651 + }, + { + "path": "1127765.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "folk", + "choice_b": "hiphop", + "choice_c": "chillout", + "choice_d": "orchestral", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20652 + }, + { + "path": "196306.mp3", + "question": "What's the flavor of music in this recording?", + "choice_a": "idm", + "choice_b": "punkrock", + "choice_c": "reggae", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20653 + }, + { + "path": "1105461.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "poprock", + "choice_b": "rnb", + "choice_c": "blues", + "choice_d": "celtic", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20654 + }, + { + "path": "189590.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "rock", + "choice_b": "soundtrack", + "choice_c": "triphop", + "choice_d": "idm", + "answer_gt": "idm", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20655 + }, + { + "path": "149054.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "electronica", + "choice_b": "blues", + "choice_c": "rock", + "choice_d": "idm", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20656 + }, + { + "path": "1127772.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "instrumentalpop", + "choice_b": "hiphop", + "choice_c": "latin", + "choice_d": "techno", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20657 + }, + { + "path": "1037375.mp3", + "question": "detect the genre of this music?", + "choice_a": "alternative", + "choice_b": "soul", + "choice_c": "soundtrack", + "choice_d": "electronica", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20658 + }, + { + "path": "965445.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "classical", + "choice_b": "orchestral", + "choice_c": "ethno", + "choice_d": "latin", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20659 + }, + { + "path": "163770.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "rap", + "choice_b": "punkrock", + "choice_c": "electronic", + "choice_d": "rnb", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20660 + }, + { + "path": "594841.mp3", + "question": "classify the genre of this music?", + "choice_a": "jazz", + "choice_b": "soundtrack", + "choice_c": "soul", + "choice_d": "classical", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20661 + }, + { + "path": "84497.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "poprock", + "choice_b": "reggae", + "choice_c": "acidjazz", + "choice_d": "darkambient", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20662 + }, + { + "path": "324380.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "easylistening", + "choice_b": "ethno", + "choice_c": "rnb", + "choice_d": "folk", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20663 + }, + { + "path": "1350314.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "newage", + "choice_b": "soundtrack", + "choice_c": "world", + "choice_d": "acidjazz", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20664 + }, + { + "path": "55634.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "rap", + "choice_b": "indie", + "choice_c": "dance", + "choice_d": "experimental", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20665 + }, + { + "path": "1174216.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "easylistening", + "choice_b": "house", + "choice_c": "world", + "choice_d": "celtic", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20666 + }, + { + "path": "592446.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "acidjazz", + "choice_b": "rock", + "choice_c": "instrumentalpop", + "choice_d": "techno", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20667 + }, + { + "path": "969349.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "blues", + "choice_b": "orchestral", + "choice_c": "classical", + "choice_d": "rnb", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20668 + }, + { + "path": "322097.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "world", + "choice_b": "celtic", + "choice_c": "electronic", + "choice_d": "hardrock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20669 + }, + { + "path": "692791.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "hiphop", + "choice_b": "rocknroll", + "choice_c": "progressive", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20670 + }, + { + "path": "1402160.mp3", + "question": "identify the music genre of this track.", + "choice_a": "electronica", + "choice_b": "ambient", + "choice_c": "blues", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20671 + }, + { + "path": "16101.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "world", + "choice_b": "soundtrack", + "choice_c": "electronic", + "choice_d": "metal", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20672 + }, + { + "path": "82884.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "newage", + "choice_b": "electropop", + "choice_c": "electronic", + "choice_d": "world", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20673 + }, + { + "path": "903592.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "classical", + "choice_b": "celtic", + "choice_c": "experimental", + "choice_d": "soul", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20674 + }, + { + "path": "285355.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "acidjazz", + "choice_b": "electronic", + "choice_c": "hardrock", + "choice_d": "pop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20675 + }, + { + "path": "1117704.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "darkambient", + "choice_b": "reggae", + "choice_c": "psychedelic", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20676 + }, + { + "path": "240488.mp3", + "question": "what's the genre of this music?", + "choice_a": "pop", + "choice_b": "electronic", + "choice_c": "bossanova", + "choice_d": "dance", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20677 + }, + { + "path": "1037370.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "breakbeat", + "choice_b": "electronica", + "choice_c": "soundtrack", + "choice_d": "techno", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20678 + }, + { + "path": "458686.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "soundtrack", + "choice_b": "jazz", + "choice_c": "chillout", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20679 + }, + { + "path": "1348924.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "darkambient", + "choice_b": "poprock", + "choice_c": "hardrock", + "choice_d": "electropop", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20680 + }, + { + "path": "1158367.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "hiphop", + "choice_b": "soundtrack", + "choice_c": "latin", + "choice_d": "darkambient", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20681 + }, + { + "path": "1399553.mp3", + "question": "classify the genre of this music?", + "choice_a": "metal", + "choice_b": "classical", + "choice_c": "hardrock", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20682 + }, + { + "path": "211828.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "minimal", + "choice_b": "ambient", + "choice_c": "country", + "choice_d": "funk", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20683 + }, + { + "path": "589553.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "ethno", + "choice_b": "classical", + "choice_c": "psychedelic", + "choice_d": "darkambient", + "answer_gt": "psychedelic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20684 + }, + { + "path": "965309.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "classical", + "choice_b": "bossanova", + "choice_c": "jazz", + "choice_d": "progressive", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20685 + }, + { + "path": "1350988.mp3", + "question": "What style of music is this?", + "choice_a": "hardrock", + "choice_b": "soundtrack", + "choice_c": "industrial", + "choice_d": "psychedelic", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20686 + }, + { + "path": "858536.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "idm", + "choice_b": "alternative", + "choice_c": "celtic", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20687 + }, + { + "path": "1085878.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "orchestral", + "choice_b": "ambient", + "choice_c": "classical", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20688 + }, + { + "path": "1154295.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "drumnbass", + "choice_b": "ambient", + "choice_c": "hardrock", + "choice_d": "indie", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20689 + }, + { + "path": "283809.mp3", + "question": "What's the classification of this music?", + "choice_a": "electronic", + "choice_b": "folk", + "choice_c": "newwave", + "choice_d": "soundtrack", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20690 + }, + { + "path": "199569.mp3", + "question": "What's the musical classification of this piece?", + "choice_a": "triphop", + "choice_b": "soul", + "choice_c": "techno", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20691 + }, + { + "path": "1214580.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "bossanova", + "choice_b": "orchestral", + "choice_c": "alternative", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20692 + }, + { + "path": "979854.mp3", + "question": "What style of music is this?", + "choice_a": "ambient", + "choice_b": "progressive", + "choice_c": "ethno", + "choice_d": "pop", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20693 + }, + { + "path": "44433.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "psychedelic", + "choice_b": "rock", + "choice_c": "experimental", + "choice_d": "easylistening", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20694 + }, + { + "path": "1354589.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "folk", + "choice_b": "soundtrack", + "choice_c": "hard", + "choice_d": "popfolk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20695 + }, + { + "path": "508156.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "easylistening", + "choice_b": "jazz", + "choice_c": "rock", + "choice_d": "breakbeat", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20696 + }, + { + "path": "1127780.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "electropop", + "choice_b": "blues", + "choice_c": "rocknroll", + "choice_d": "hiphop", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20697 + }, + { + "path": "1187041.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "blues", + "choice_b": "minimal", + "choice_c": "electropop", + "choice_d": "orchestral", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20698 + }, + { + "path": "1117422.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "classical", + "choice_b": "drumnbass", + "choice_c": "acidjazz", + "choice_d": "metal", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20699 + }, + { + "path": "1245902.mp3", + "question": "In what genre would you place this tune?", + "choice_a": "rock", + "choice_b": "acidjazz", + "choice_c": "orchestral", + "choice_d": "house", + "answer_gt": "acidjazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20700 + }, + { + "path": "1353563.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "electronic", + "choice_b": "soundtrack", + "choice_c": "hard", + "choice_d": "funk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20701 + }, + { + "path": "1135221.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "alternative", + "choice_b": "newage", + "choice_c": "folk", + "choice_d": "techno", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20702 + }, + { + "path": "179543.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "newwave", + "choice_b": "minimal", + "choice_c": "hiphop", + "choice_d": "soul", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20703 + }, + { + "path": "285659.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "hard", + "choice_b": "house", + "choice_c": "classical", + "choice_d": "alternative", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20704 + }, + { + "path": "270117.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "indie", + "choice_b": "darkambient", + "choice_c": "classical", + "choice_d": "funk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20705 + }, + { + "path": "1349693.mp3", + "question": "what's the genre of this music?", + "choice_a": "techno", + "choice_b": "rock", + "choice_c": "soundtrack", + "choice_d": "popfolk", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20706 + }, + { + "path": "285621.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "darkambient", + "choice_b": "classical", + "choice_c": "atmospheric", + "choice_d": "acidjazz", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20707 + }, + { + "path": "1286488.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "easylistening", + "choice_b": "celtic", + "choice_c": "folk", + "choice_d": "newage", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20708 + }, + { + "path": "1296780.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "blues", + "choice_b": "swing", + "choice_c": "electronica", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20709 + }, + { + "path": "1286493.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "triphop", + "choice_b": "rnb", + "choice_c": "newage", + "choice_d": "rocknroll", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20710 + }, + { + "path": "908060.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electropop", + "choice_b": "soundtrack", + "choice_c": "easylistening", + "choice_d": "triphop", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20711 + }, + { + "path": "1119926.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "blues", + "choice_b": "rock", + "choice_c": "soundtrack", + "choice_d": "country", + "answer_gt": "country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20712 + }, + { + "path": "368182.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "funk", + "choice_b": "classical", + "choice_c": "jazz", + "choice_d": "house", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20713 + }, + { + "path": "1238881.mp3", + "question": "What style of music is this?", + "choice_a": "electropop", + "choice_b": "orchestral", + "choice_c": "electronica", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20714 + }, + { + "path": "36983.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "indie", + "choice_b": "techno", + "choice_c": "atmospheric", + "choice_d": "world", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20715 + }, + { + "path": "873131.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "classical", + "choice_b": "world", + "choice_c": "minimal", + "choice_d": "alternative", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20716 + }, + { + "path": "979845.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "poprock", + "choice_b": "techno", + "choice_c": "ambient", + "choice_d": "indie", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20717 + }, + { + "path": "434896.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "chanson", + "choice_b": "pop", + "choice_c": "country", + "choice_d": "hard", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20718 + }, + { + "path": "1159259.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "techno", + "choice_b": "pop", + "choice_c": "folk", + "choice_d": "electropop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20719 + }, + { + "path": "1076577.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "hiphop", + "choice_b": "techno", + "choice_c": "hardrock", + "choice_d": "atmospheric", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20720 + }, + { + "path": "1372934.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "instrumentalpop", + "choice_b": "ethno", + "choice_c": "rock", + "choice_d": "electronic", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20721 + }, + { + "path": "1128296.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "newwave", + "choice_b": "punkrock", + "choice_c": "house", + "choice_d": "celtic", + "answer_gt": "celtic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20722 + }, + { + "path": "944154.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "breakbeat", + "choice_b": "alternative", + "choice_c": "soul", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20723 + }, + { + "path": "324387.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "dance", + "choice_b": "classical", + "choice_c": "hiphop", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20724 + }, + { + "path": "1075261.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "rap", + "choice_b": "popfolk", + "choice_c": "pop", + "choice_d": "chillout", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20725 + }, + { + "path": "490363.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "poprock", + "choice_b": "techno", + "choice_c": "folk", + "choice_d": "latin", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20726 + }, + { + "path": "18500.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "breakbeat", + "choice_b": "electronic", + "choice_c": "celtic", + "choice_d": "electropop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20727 + }, + { + "path": "965307.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "classical", + "choice_b": "industrial", + "choice_c": "popfolk", + "choice_d": "folk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20728 + }, + { + "path": "1350821.mp3", + "question": "specify the music genre here?", + "choice_a": "newwave", + "choice_b": "soundtrack", + "choice_c": "world", + "choice_d": "trance", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20729 + }, + { + "path": "224793.mp3", + "question": "What genre does this tune fall under?", + "choice_a": "electronic", + "choice_b": "popfolk", + "choice_c": "triphop", + "choice_d": "easylistening", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20730 + }, + { + "path": "1348849.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "poprock", + "choice_b": "orchestral", + "choice_c": "dance", + "choice_d": "popfolk", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20731 + }, + { + "path": "201991.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "orchestral", + "choice_b": "jazz", + "choice_c": "newwave", + "choice_d": "techno", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20732 + }, + { + "path": "278149.mp3", + "question": "How would you describe the music genre here?", + "choice_a": "metal", + "choice_b": "classical", + "choice_c": "jazz", + "choice_d": "chillout", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20733 + }, + { + "path": "347284.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "ambient", + "choice_b": "rock", + "choice_c": "electropop", + "choice_d": "classical", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20734 + }, + { + "path": "1072136.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "industrial", + "choice_b": "poprock", + "choice_c": "techno", + "choice_d": "funk", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20735 + }, + { + "path": "1286512.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "indie", + "choice_b": "chanson", + "choice_c": "newage", + "choice_d": "funk", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20736 + }, + { + "path": "686006.mp3", + "question": "What's the classification of this music?", + "choice_a": "electronic", + "choice_b": "orchestral", + "choice_c": "popfolk", + "choice_d": "rnb", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20737 + }, + { + "path": "1076581.mp3", + "question": "What's the genre feel of this track?", + "choice_a": "electropop", + "choice_b": "hiphop", + "choice_c": "reggae", + "choice_d": "ambient", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20738 + }, + { + "path": "1348315.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "reggae", + "choice_b": "soundtrack", + "choice_c": "progressive", + "choice_d": "swing", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20739 + }, + { + "path": "347202.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "rap", + "choice_b": "experimental", + "choice_c": "rock", + "choice_d": "dance", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20740 + }, + { + "path": "199280.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "electronica", + "choice_b": "breakbeat", + "choice_c": "folk", + "choice_d": "soul", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20741 + }, + { + "path": "1109777.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "blues", + "choice_b": "breakbeat", + "choice_c": "electronic", + "choice_d": "rock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20742 + }, + { + "path": "965447.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "house", + "choice_b": "classical", + "choice_c": "easylistening", + "choice_d": "experimental", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20743 + }, + { + "path": "208192.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "swing", + "choice_b": "rock", + "choice_c": "easylistening", + "choice_d": "chanson", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20744 + }, + { + "path": "1286509.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "newage", + "choice_b": "rap", + "choice_c": "techno", + "choice_d": "chillout", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20745 + }, + { + "path": "1026042.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "jazz", + "choice_b": "progressive", + "choice_c": "bossanova", + "choice_d": "house", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20746 + }, + { + "path": "129557.mp3", + "question": "What style of music is this?", + "choice_a": "reggae", + "choice_b": "progressive", + "choice_c": "alternative", + "choice_d": "hardrock", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20747 + }, + { + "path": "1034298.mp3", + "question": "What's the classification of this music?", + "choice_a": "chanson", + "choice_b": "jazz", + "choice_c": "pop", + "choice_d": "rock", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20748 + }, + { + "path": "1100153.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "indie", + "choice_b": "pop", + "choice_c": "metal", + "choice_d": "alternative", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20749 + }, + { + "path": "1362286.mp3", + "question": "How would you interpret the genre of this song?", + "choice_a": "folk", + "choice_b": "dance", + "choice_c": "hard", + "choice_d": "hardrock", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20750 + }, + { + "path": "172025.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "hardrock", + "choice_b": "world", + "choice_c": "hiphop", + "choice_d": "newwave", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20751 + }, + { + "path": "217496.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "electronic", + "choice_b": "swing", + "choice_c": "reggae", + "choice_d": "indie", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20752 + }, + { + "path": "847262.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "classical", + "choice_b": "electropop", + "choice_c": "dance", + "choice_d": "rock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20753 + }, + { + "path": "1210875.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "bossanova", + "choice_b": "pop", + "choice_c": "world", + "choice_d": "country", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20754 + }, + { + "path": "42345.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "rock", + "choice_b": "swing", + "choice_c": "reggae", + "choice_d": "triphop", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20755 + }, + { + "path": "1352578.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "psychedelic", + "choice_b": "soundtrack", + "choice_c": "progressive", + "choice_d": "alternative", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20756 + }, + { + "path": "970790.mp3", + "question": "what's the genre of this music?", + "choice_a": "idm", + "choice_b": "rocknroll", + "choice_c": "trance", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20757 + }, + { + "path": "1086828.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "techno", + "choice_b": "rnb", + "choice_c": "classical", + "choice_d": "reggae", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20758 + }, + { + "path": "1062501.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "ambient", + "choice_b": "classical", + "choice_c": "soundtrack", + "choice_d": "world", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20759 + }, + { + "path": "1394724.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "alternative", + "choice_b": "electronica", + "choice_c": "chanson", + "choice_d": "pop", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20760 + }, + { + "path": "243835.mp3", + "question": "What's the genre signature of this song?", + "choice_a": "psychedelic", + "choice_b": "atmospheric", + "choice_c": "progressive", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20761 + }, + { + "path": "1347935.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "hardrock", + "choice_b": "pop", + "choice_c": "electronic", + "choice_d": "rap", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20762 + }, + { + "path": "1077910.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "newage", + "choice_b": "techno", + "choice_c": "orchestral", + "choice_d": "pop", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20763 + }, + { + "path": "1352140.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "soundtrack", + "choice_b": "easylistening", + "choice_c": "poprock", + "choice_d": "instrumentalpop", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20764 + }, + { + "path": "90131.mp3", + "question": "Which genre realm does this music belong to?", + "choice_a": "orchestral", + "choice_b": "electronica", + "choice_c": "electropop", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20765 + }, + { + "path": "1267.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "newwave", + "choice_b": "hardrock", + "choice_c": "progressive", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20766 + }, + { + "path": "127564.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "psychedelic", + "choice_b": "electropop", + "choice_c": "hard", + "choice_d": "easylistening", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20767 + }, + { + "path": "345620.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "industrial", + "choice_b": "easylistening", + "choice_c": "house", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20768 + }, + { + "path": "15069.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "electronic", + "choice_b": "country", + "choice_c": "hardrock", + "choice_d": "funk", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20769 + }, + { + "path": "604137.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "electronic", + "choice_b": "folk", + "choice_c": "swing", + "choice_d": "world", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20770 + }, + { + "path": "129559.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "classical", + "choice_b": "alternative", + "choice_c": "bossanova", + "choice_d": "electronica", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20771 + }, + { + "path": "1133077.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "punkrock", + "choice_b": "folk", + "choice_c": "classical", + "choice_d": "drumnbass", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20772 + }, + { + "path": "441084.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "experimental", + "choice_b": "drumnbass", + "choice_c": "country", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20773 + }, + { + "path": "156970.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "progressive", + "choice_b": "hiphop", + "choice_c": "dance", + "choice_d": "celtic", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20774 + }, + { + "path": "1391798.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "soul", + "choice_b": "bossanova", + "choice_c": "easylistening", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20775 + }, + { + "path": "1119041.mp3", + "question": "Which musical genre is this composition associated with?", + "choice_a": "idm", + "choice_b": "electronic", + "choice_c": "industrial", + "choice_d": "electronica", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20776 + }, + { + "path": "434893.mp3", + "question": "specify the music genre here?", + "choice_a": "chanson", + "choice_b": "electropop", + "choice_c": "jazz", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20777 + }, + { + "path": "179536.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "jazz", + "choice_b": "easylistening", + "choice_c": "hiphop", + "choice_d": "experimental", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20778 + }, + { + "path": "1127785.mp3", + "question": "specify the music genre here?", + "choice_a": "funk", + "choice_b": "country", + "choice_c": "hiphop", + "choice_d": "easylistening", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20779 + }, + { + "path": "253718.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "rock", + "choice_b": "blues", + "choice_c": "progressive", + "choice_d": "rnb", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20780 + }, + { + "path": "17623.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "chanson", + "choice_b": "reggae", + "choice_c": "idm", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20781 + }, + { + "path": "211843.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "rocknroll", + "choice_b": "chanson", + "choice_c": "electronic", + "choice_d": "classical", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20782 + }, + { + "path": "132153.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "reggae", + "choice_b": "country", + "choice_c": "soul", + "choice_d": "alternative", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20783 + }, + { + "path": "1350956.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "folk", + "choice_b": "soundtrack", + "choice_c": "psychedelic", + "choice_d": "progressive", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20784 + }, + { + "path": "664821.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "rocknroll", + "choice_b": "classical", + "choice_c": "soul", + "choice_d": "hard", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20785 + }, + { + "path": "1173330.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "electropop", + "choice_b": "alternative", + "choice_c": "swing", + "choice_d": "blues", + "answer_gt": "alternative", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20786 + }, + { + "path": "1279181.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "trance", + "choice_b": "world", + "choice_c": "newwave", + "choice_d": "easylistening", + "answer_gt": "world", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20787 + }, + { + "path": "149490.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "classical", + "choice_b": "newage", + "choice_c": "industrial", + "choice_d": "bossanova", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20788 + }, + { + "path": "1100241.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "techno", + "choice_b": "popfolk", + "choice_c": "pop", + "choice_d": "hardrock", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20789 + }, + { + "path": "972768.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "metal", + "choice_b": "atmospheric", + "choice_c": "minimal", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20790 + }, + { + "path": "1088582.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "jazz", + "choice_b": "latin", + "choice_c": "triphop", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20791 + }, + { + "path": "1007829.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "drumnbass", + "choice_b": "rap", + "choice_c": "jazz", + "choice_d": "poprock", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20792 + }, + { + "path": "163768.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "alternative", + "choice_b": "latin", + "choice_c": "hiphop", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20793 + }, + { + "path": "1350953.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "punkrock", + "choice_b": "chanson", + "choice_c": "soundtrack", + "choice_d": "electropop", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20794 + }, + { + "path": "1076579.mp3", + "question": "specify the music genre here?", + "choice_a": "hiphop", + "choice_b": "electropop", + "choice_c": "drumnbass", + "choice_d": "house", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20795 + }, + { + "path": "616358.mp3", + "question": "Which genre would you say this piece of music fits into?", + "choice_a": "newage", + "choice_b": "jazz", + "choice_c": "chillout", + "choice_d": "hiphop", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20796 + }, + { + "path": "1150898.mp3", + "question": "what's the genre of this music?", + "choice_a": "industrial", + "choice_b": "rock", + "choice_c": "blues", + "choice_d": "pop", + "answer_gt": "blues", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20797 + }, + { + "path": "768502.mp3", + "question": "Which genre box does this track tick?", + "choice_a": "drumnbass", + "choice_b": "hardrock", + "choice_c": "electronic", + "choice_d": "house", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20798 + }, + { + "path": "1067233.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "electronic", + "choice_b": "classical", + "choice_c": "blues", + "choice_d": "rnb", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20799 + }, + { + "path": "1088593.mp3", + "question": "classify the genre of this music?", + "choice_a": "breakbeat", + "choice_b": "celtic", + "choice_c": "popfolk", + "choice_d": "pop", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20800 + }, + { + "path": "701770.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "triphop", + "choice_b": "electronic", + "choice_c": "hard", + "choice_d": "ethno", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20801 + }, + { + "path": "1116267.mp3", + "question": "What kind of music is playing in this audio?", + "choice_a": "rocknroll", + "choice_b": "alternative", + "choice_c": "country", + "choice_d": "instrumentalpop", + "answer_gt": "country", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20802 + }, + { + "path": "1090462.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "minimal", + "choice_b": "pop", + "choice_c": "rap", + "choice_d": "soundtrack", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20803 + }, + { + "path": "295884.mp3", + "question": "What genre of music is featured in this recording?", + "choice_a": "indie", + "choice_b": "hardrock", + "choice_c": "rock", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20804 + }, + { + "path": "147713.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "swing", + "choice_b": "classical", + "choice_c": "soul", + "choice_d": "blues", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20805 + }, + { + "path": "1160010.mp3", + "question": "What genre does this musical piece exemplify?", + "choice_a": "folk", + "choice_b": "trance", + "choice_c": "bossanova", + "choice_d": "house", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20806 + }, + { + "path": "1350668.mp3", + "question": "How would you summarize the genre of this music?", + "choice_a": "poprock", + "choice_b": "punkrock", + "choice_c": "breakbeat", + "choice_d": "triphop", + "answer_gt": "poprock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20807 + }, + { + "path": "253749.mp3", + "question": "What's the genre identity of this music?", + "choice_a": "electronic", + "choice_b": "rock", + "choice_c": "ambient", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20808 + }, + { + "path": "965761.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "chillout", + "choice_b": "jazz", + "choice_c": "electropop", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20809 + }, + { + "path": "1003073.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "metal", + "choice_b": "rap", + "choice_c": "jazz", + "choice_d": "drumnbass", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20810 + }, + { + "path": "179686.mp3", + "question": "What's the genre tone of this track?", + "choice_a": "blues", + "choice_b": "hiphop", + "choice_c": "soundtrack", + "choice_d": "celtic", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20811 + }, + { + "path": "276974.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "classical", + "choice_b": "hard", + "choice_c": "dance", + "choice_d": "easylistening", + "answer_gt": "dance", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20812 + }, + { + "path": "1109358.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "jazz", + "choice_b": "bossanova", + "choice_c": "techno", + "choice_d": "chillout", + "answer_gt": "techno", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20813 + }, + { + "path": "284415.mp3", + "question": "detect the genre of this music?", + "choice_a": "rap", + "choice_b": "electronic", + "choice_c": "jazz", + "choice_d": "chillout", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20814 + }, + { + "path": "914773.mp3", + "question": "What's the sound genre of this audio clip?", + "choice_a": "funk", + "choice_b": "electropop", + "choice_c": "folk", + "choice_d": "house", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20815 + }, + { + "path": "1149757.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "classical", + "choice_b": "hiphop", + "choice_c": "experimental", + "choice_d": "pop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20816 + }, + { + "path": "1086430.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "reggae", + "choice_b": "hardrock", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20817 + }, + { + "path": "42340.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "bossanova", + "choice_b": "triphop", + "choice_c": "atmospheric", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20818 + }, + { + "path": "1075326.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "instrumentalpop", + "choice_b": "house", + "choice_c": "pop", + "choice_d": "poprock", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20819 + }, + { + "path": "184197.mp3", + "question": "specify the music genre here?", + "choice_a": "newage", + "choice_b": "metal", + "choice_c": "darkambient", + "choice_d": "techno", + "answer_gt": "newage", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20820 + }, + { + "path": "896186.mp3", + "question": "What's the classification of this music?", + "choice_a": "electronic", + "choice_b": "drumnbass", + "choice_c": "punkrock", + "choice_d": "psychedelic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20821 + }, + { + "path": "847329.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "electronic", + "choice_b": "idm", + "choice_c": "country", + "choice_d": "electropop", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20822 + }, + { + "path": "1350133.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "country", + "choice_b": "soundtrack", + "choice_c": "folk", + "choice_d": "experimental", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20823 + }, + { + "path": "1116232.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "atmospheric", + "choice_b": "latin", + "choice_c": "newage", + "choice_d": "darkambient", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20824 + }, + { + "path": "814947.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "industrial", + "choice_b": "breakbeat", + "choice_c": "drumnbass", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20825 + }, + { + "path": "133097.mp3", + "question": "what's the genre of this music?", + "choice_a": "electronic", + "choice_b": "indie", + "choice_c": "hiphop", + "choice_d": "bossanova", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20826 + }, + { + "path": "502332.mp3", + "question": "How would you define the genre of this music?", + "choice_a": "electronic", + "choice_b": "classical", + "choice_c": "orchestral", + "choice_d": "funk", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20827 + }, + { + "path": "944147.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "blues", + "choice_b": "classical", + "choice_c": "ethno", + "choice_d": "pop", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20828 + }, + { + "path": "862194.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "jazz", + "choice_b": "chillout", + "choice_c": "classical", + "choice_d": "rocknroll", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20829 + }, + { + "path": "1352149.mp3", + "question": "identify the music genre of this track.", + "choice_a": "rap", + "choice_b": "electronica", + "choice_c": "orchestral", + "choice_d": "soundtrack", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20830 + }, + { + "path": "587000.mp3", + "question": "detect the genre of this music?", + "choice_a": "atmospheric", + "choice_b": "soundtrack", + "choice_c": "electronica", + "choice_d": "rap", + "answer_gt": "atmospheric", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20831 + }, + { + "path": "106895.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "reggae", + "choice_b": "electronic", + "choice_c": "idm", + "choice_d": "indie", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20832 + }, + { + "path": "1350989.mp3", + "question": "classify the genre of this music?", + "choice_a": "hard", + "choice_b": "newwave", + "choice_c": "poprock", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20833 + }, + { + "path": "893495.mp3", + "question": "tell me the genre of this music piece?", + "choice_a": "industrial", + "choice_b": "rock", + "choice_c": "easylistening", + "choice_d": "reggae", + "answer_gt": "easylistening", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20834 + }, + { + "path": "1334.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "newage", + "choice_b": "funk", + "choice_c": "electronic", + "choice_d": "poprock", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20835 + }, + { + "path": "963305.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "jazz", + "choice_b": "minimal", + "choice_c": "popfolk", + "choice_d": "classical", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20836 + }, + { + "path": "1055021.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "metal", + "choice_b": "house", + "choice_c": "experimental", + "choice_d": "electronic", + "answer_gt": "metal", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20837 + }, + { + "path": "814952.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "swing", + "choice_b": "electropop", + "choice_c": "reggae", + "choice_d": "alternative", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20838 + }, + { + "path": "401451.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "country", + "choice_b": "electronic", + "choice_c": "orchestral", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20839 + }, + { + "path": "179539.mp3", + "question": "What kind of musical genre is this?", + "choice_a": "soul", + "choice_b": "hiphop", + "choice_c": "instrumentalpop", + "choice_d": "reggae", + "answer_gt": "hiphop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20840 + }, + { + "path": "1119038.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "alternative", + "choice_b": "chillout", + "choice_c": "hard", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20841 + }, + { + "path": "967568.mp3", + "question": "What's the musical theme genre of this song?", + "choice_a": "classical", + "choice_b": "soundtrack", + "choice_c": "funk", + "choice_d": "chanson", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20842 + }, + { + "path": "1269920.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "industrial", + "choice_b": "country", + "choice_c": "rnb", + "choice_d": "instrumentalpop", + "answer_gt": "industrial", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20843 + }, + { + "path": "253720.mp3", + "question": "What musical category does this song belong to?", + "choice_a": "soundtrack", + "choice_b": "latin", + "choice_c": "electronic", + "choice_d": "dance", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20844 + }, + { + "path": "1117701.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "metal", + "choice_b": "minimal", + "choice_c": "classical", + "choice_d": "punkrock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20845 + }, + { + "path": "246548.mp3", + "question": "Which music genre does this track seem to be?", + "choice_a": "ambient", + "choice_b": "electronic", + "choice_c": "rocknroll", + "choice_d": "classical", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20846 + }, + { + "path": "1215849.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "psychedelic", + "choice_b": "easylistening", + "choice_c": "rocknroll", + "choice_d": "latin", + "answer_gt": "latin", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20847 + }, + { + "path": "253719.mp3", + "question": "What's the predominant music genre in this track?", + "choice_a": "swing", + "choice_b": "ethno", + "choice_c": "alternative", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20848 + }, + { + "path": "1351468.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "chillout", + "choice_b": "electropop", + "choice_c": "soundtrack", + "choice_d": "dance", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20849 + }, + { + "path": "979850.mp3", + "question": "What genre vibe does this music give off?", + "choice_a": "soul", + "choice_b": "indie", + "choice_c": "poprock", + "choice_d": "ambient", + "answer_gt": "ambient", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20850 + }, + { + "path": "1173976.mp3", + "question": "identify the music genre of this track.", + "choice_a": "drumnbass", + "choice_b": "house", + "choice_c": "triphop", + "choice_d": "punkrock", + "answer_gt": "house", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20851 + }, + { + "path": "82883.mp3", + "question": "What's the genre mood of this music?", + "choice_a": "popfolk", + "choice_b": "house", + "choice_c": "instrumentalpop", + "choice_d": "electronic", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20852 + }, + { + "path": "1415796.mp3", + "question": "What kind of genre is this music associated with?", + "choice_a": "electronic", + "choice_b": "ambient", + "choice_c": "easylistening", + "choice_d": "metal", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20853 + }, + { + "path": "879146.mp3", + "question": "specify the music genre here?", + "choice_a": "metal", + "choice_b": "ethno", + "choice_c": "chillout", + "choice_d": "world", + "answer_gt": "chillout", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20854 + }, + { + "path": "694397.mp3", + "question": "How would you label the genre of this track?", + "choice_a": "hardrock", + "choice_b": "atmospheric", + "choice_c": "punkrock", + "choice_d": "progressive", + "answer_gt": "punkrock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20855 + }, + { + "path": "1159863.mp3", + "question": "What's the genre essence of this musical piece?", + "choice_a": "psychedelic", + "choice_b": "indie", + "choice_c": "breakbeat", + "choice_d": "country", + "answer_gt": "indie", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20856 + }, + { + "path": "1054368.mp3", + "question": "What's the genre atmosphere of this piece?", + "choice_a": "pop", + "choice_b": "classical", + "choice_c": "newwave", + "choice_d": "hardrock", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20857 + }, + { + "path": "393259.mp3", + "question": "What's the stylistic genre of this track?", + "choice_a": "triphop", + "choice_b": "alternative", + "choice_c": "soundtrack", + "choice_d": "classical", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20858 + }, + { + "path": "211832.mp3", + "question": "detect the genre of this music?", + "choice_a": "jazz", + "choice_b": "poprock", + "choice_c": "atmospheric", + "choice_d": "triphop", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20859 + }, + { + "path": "1350142.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "alternative", + "choice_b": "soundtrack", + "choice_c": "reggae", + "choice_d": "house", + "answer_gt": "soundtrack", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20860 + }, + { + "path": "1173985.mp3", + "question": "What's the genre theme of this audio clip?", + "choice_a": "rap", + "choice_b": "electronic", + "choice_c": "house", + "choice_d": "popfolk", + "answer_gt": "house", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20861 + }, + { + "path": "1133073.mp3", + "question": "specify the music genre here?", + "choice_a": "rocknroll", + "choice_b": "classical", + "choice_c": "hiphop", + "choice_d": "country", + "answer_gt": "classical", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20862 + }, + { + "path": "575226.mp3", + "question": "detect the genre of this music?", + "choice_a": "dance", + "choice_b": "folk", + "choice_c": "darkambient", + "choice_d": "hardrock", + "answer_gt": "folk", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20863 + }, + { + "path": "639280.mp3", + "question": "What family of music does this piece come from?", + "choice_a": "acidjazz", + "choice_b": "electronic", + "choice_c": "ambient", + "choice_d": "alternative", + "answer_gt": "electronic", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20864 + }, + { + "path": "760.mp3", + "question": "pinpoint the music genre of this piece?", + "choice_a": "pop", + "choice_b": "psychedelic", + "choice_c": "house", + "choice_d": "minimal", + "answer_gt": "pop", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20865 + }, + { + "path": "876296.mp3", + "question": "How would you categorize the genre of this music?", + "choice_a": "darkambient", + "choice_b": "rocknroll", + "choice_c": "jazz", + "choice_d": "soundtrack", + "answer_gt": "jazz", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20866 + }, + { + "path": "1305896.mp3", + "question": "What's the type of music we're listening to here?", + "choice_a": "breakbeat", + "choice_b": "hiphop", + "choice_c": "soundtrack", + "choice_d": "reggae", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20867 + }, + { + "path": "196129.mp3", + "question": "What style of music is this?", + "choice_a": "rock", + "choice_b": "house", + "choice_c": "classical", + "choice_d": "reggae", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20868 + }, + { + "path": "381708.mp3", + "question": "What's the musical genre of this song?", + "choice_a": "reggae", + "choice_b": "poprock", + "choice_c": "industrial", + "choice_d": "country", + "answer_gt": "reggae", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20869 + }, + { + "path": "199623.mp3", + "question": "Which genre does this piece of music represent?", + "choice_a": "atmospheric", + "choice_b": "country", + "choice_c": "classical", + "choice_d": "rock", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20870 + }, + { + "path": "1051717.mp3", + "question": "assess the music genre of this recording?", + "choice_a": "punkrock", + "choice_b": "rock", + "choice_c": "newage", + "choice_d": "country", + "answer_gt": "rock", + "task_name": "Music_Genre_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 20871 + }, + { + "path": "flute_acoustic_002-103-050.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20872 + }, + { + "path": "bass_electronic_025-059-050.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20873 + }, + { + "path": "keyboard_electronic_002-033-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20874 + }, + { + "path": "guitar_acoustic_014-067-050.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20875 + }, + { + "path": "keyboard_synthetic_000-063-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20876 + }, + { + "path": "guitar_acoustic_030-064-127.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20877 + }, + { + "path": "bass_synthetic_009-044-050.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20878 + }, + { + "path": "bass_synthetic_033-105-050.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_105", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20879 + }, + { + "path": "keyboard_electronic_078-061-025.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20880 + }, + { + "path": "guitar_acoustic_010-027-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20881 + }, + { + "path": "mallet_acoustic_062-107-075.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_13", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20882 + }, + { + "path": "keyboard_electronic_001-037-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20883 + }, + { + "path": "brass_acoustic_015-045-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20884 + }, + { + "path": "mallet_acoustic_056-088-075.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20885 + }, + { + "path": "bass_electronic_018-043-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20886 + }, + { + "path": "flute_synthetic_000-101-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20887 + }, + { + "path": "organ_electronic_007-096-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20888 + }, + { + "path": "organ_electronic_057-050-025.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20889 + }, + { + "path": "keyboard_electronic_003-047-025.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20890 + }, + { + "path": "reed_acoustic_018-072-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20891 + }, + { + "path": "bass_synthetic_068-084-127.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_20", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20892 + }, + { + "path": "keyboard_electronic_069-068-075.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20893 + }, + { + "path": "bass_synthetic_033-065-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20894 + }, + { + "path": "keyboard_electronic_001-024-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20895 + }, + { + "path": "organ_electronic_028-068-100.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20896 + }, + { + "path": "keyboard_electronic_078-082-100.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20897 + }, + { + "path": "vocal_synthetic_003-081-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20898 + }, + { + "path": "organ_electronic_057-025-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20899 + }, + { + "path": "keyboard_electronic_001-107-025.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20900 + }, + { + "path": "organ_electronic_007-038-075.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20901 + }, + { + "path": "bass_synthetic_134-096-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20902 + }, + { + "path": "reed_acoustic_023-097-100.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_97", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20903 + }, + { + "path": "guitar_electronic_028-056-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20904 + }, + { + "path": "organ_electronic_007-012-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_12", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20905 + }, + { + "path": "flute_synthetic_000-037-075.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20906 + }, + { + "path": "reed_acoustic_023-054-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20907 + }, + { + "path": "bass_synthetic_098-021-100.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20908 + }, + { + "path": "flute_synthetic_000-096-050.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20909 + }, + { + "path": "organ_electronic_001-058-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20910 + }, + { + "path": "guitar_acoustic_021-055-100.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20911 + }, + { + "path": "keyboard_electronic_069-052-025.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20912 + }, + { + "path": "mallet_acoustic_047-085-050.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_16", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20913 + }, + { + "path": "vocal_synthetic_003-096-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20914 + }, + { + "path": "keyboard_acoustic_004-079-025.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_40", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20915 + }, + { + "path": "brass_acoustic_006-032-127.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20916 + }, + { + "path": "keyboard_acoustic_004-049-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20917 + }, + { + "path": "keyboard_acoustic_004-025-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_28", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20918 + }, + { + "path": "bass_synthetic_009-033-075.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20919 + }, + { + "path": "guitar_electronic_022-053-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20920 + }, + { + "path": "organ_electronic_028-106-050.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_40", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20921 + }, + { + "path": "organ_electronic_057-068-127.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20922 + }, + { + "path": "reed_acoustic_037-073-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20923 + }, + { + "path": "string_acoustic_080-034-127.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20924 + }, + { + "path": "brass_acoustic_059-050-127.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20925 + }, + { + "path": "reed_acoustic_037-057-100.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20926 + }, + { + "path": "reed_acoustic_037-055-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_55", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_28", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20927 + }, + { + "path": "guitar_electronic_022-040-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20928 + }, + { + "path": "flute_acoustic_002-069-127.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20929 + }, + { + "path": "organ_electronic_028-081-050.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_16", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20930 + }, + { + "path": "organ_electronic_113-065-050.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20931 + }, + { + "path": "bass_electronic_025-030-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20932 + }, + { + "path": "keyboard_electronic_002-084-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20933 + }, + { + "path": "flute_acoustic_002-107-127.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20934 + }, + { + "path": "organ_electronic_028-027-127.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20935 + }, + { + "path": "guitar_acoustic_021-035-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20936 + }, + { + "path": "keyboard_synthetic_000-050-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20937 + }, + { + "path": "bass_synthetic_068-075-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20938 + }, + { + "path": "guitar_acoustic_021-040-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20939 + }, + { + "path": "organ_electronic_007-022-100.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_22", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20940 + }, + { + "path": "bass_synthetic_033-050-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20941 + }, + { + "path": "bass_synthetic_098-083-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20942 + }, + { + "path": "vocal_synthetic_003-068-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20943 + }, + { + "path": "bass_synthetic_134-039-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20944 + }, + { + "path": "keyboard_acoustic_004-103-100.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20945 + }, + { + "path": "keyboard_synthetic_000-073-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20946 + }, + { + "path": "brass_acoustic_016-084-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20947 + }, + { + "path": "string_acoustic_057-065-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20948 + }, + { + "path": "bass_synthetic_009-091-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20949 + }, + { + "path": "bass_synthetic_134-044-100.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20950 + }, + { + "path": "reed_acoustic_023-059-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20951 + }, + { + "path": "keyboard_electronic_098-034-075.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20952 + }, + { + "path": "organ_electronic_113-047-127.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20953 + }, + { + "path": "organ_electronic_057-053-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20954 + }, + { + "path": "guitar_acoustic_015-071-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_9", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20955 + }, + { + "path": "keyboard_acoustic_004-069-127.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20956 + }, + { + "path": "keyboard_electronic_003-058-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20957 + }, + { + "path": "keyboard_synthetic_000-051-100.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20958 + }, + { + "path": "string_acoustic_056-064-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20959 + }, + { + "path": "bass_synthetic_009-009-050.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_9", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20960 + }, + { + "path": "guitar_acoustic_014-043-075.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20961 + }, + { + "path": "guitar_electronic_028-050-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_113", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20962 + }, + { + "path": "string_acoustic_071-038-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20963 + }, + { + "path": "guitar_acoustic_010-043-100.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20964 + }, + { + "path": "vocal_synthetic_003-103-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20965 + }, + { + "path": "keyboard_electronic_078-069-025.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20966 + }, + { + "path": "guitar_acoustic_010-071-025.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20967 + }, + { + "path": "keyboard_electronic_002-033-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20968 + }, + { + "path": "bass_electronic_027-031-025.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20969 + }, + { + "path": "keyboard_electronic_003-088-075.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20970 + }, + { + "path": "keyboard_electronic_002-047-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20971 + }, + { + "path": "bass_synthetic_068-036-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20972 + }, + { + "path": "reed_acoustic_018-086-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20973 + }, + { + "path": "flute_acoustic_002-081-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20974 + }, + { + "path": "guitar_acoustic_014-106-100.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20975 + }, + { + "path": "keyboard_acoustic_004-056-075.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20976 + }, + { + "path": "guitar_acoustic_030-058-050.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_35", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_19", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20977 + }, + { + "path": "guitar_acoustic_015-075-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20978 + }, + { + "path": "reed_acoustic_011-057-050.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20979 + }, + { + "path": "bass_electronic_027-035-100.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20980 + }, + { + "path": "bass_synthetic_034-102-025.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_9", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_102", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20981 + }, + { + "path": "guitar_acoustic_021-082-050.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20982 + }, + { + "path": "organ_electronic_007-096-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20983 + }, + { + "path": "organ_electronic_113-045-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20984 + }, + { + "path": "guitar_acoustic_015-102-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_9", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_102", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20985 + }, + { + "path": "vocal_synthetic_003-093-100.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20986 + }, + { + "path": "keyboard_electronic_001-031-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20987 + }, + { + "path": "keyboard_electronic_003-028-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20988 + }, + { + "path": "bass_electronic_027-023-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_23", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20989 + }, + { + "path": "bass_synthetic_068-057-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20990 + }, + { + "path": "keyboard_electronic_098-023-100.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20991 + }, + { + "path": "bass_synthetic_068-038-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20992 + }, + { + "path": "keyboard_electronic_002-095-127.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_11", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20993 + }, + { + "path": "mallet_acoustic_047-089-050.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20994 + }, + { + "path": "keyboard_electronic_069-046-025.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20995 + }, + { + "path": "organ_electronic_057-072-050.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20996 + }, + { + "path": "bass_synthetic_033-044-050.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20997 + }, + { + "path": "organ_electronic_007-064-100.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20998 + }, + { + "path": "bass_synthetic_068-049-075.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 20999 + }, + { + "path": "keyboard_electronic_098-065-050.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21000 + }, + { + "path": "brass_acoustic_016-088-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21001 + }, + { + "path": "guitar_electronic_022-072-050.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21002 + }, + { + "path": "keyboard_electronic_001-034-100.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21003 + }, + { + "path": "bass_synthetic_009-034-100.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21004 + }, + { + "path": "brass_acoustic_046-071-050.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21005 + }, + { + "path": "bass_synthetic_135-035-025.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21006 + }, + { + "path": "bass_synthetic_098-045-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21007 + }, + { + "path": "guitar_acoustic_014-067-025.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21008 + }, + { + "path": "guitar_acoustic_014-073-050.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21009 + }, + { + "path": "keyboard_electronic_001-103-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21010 + }, + { + "path": "keyboard_electronic_098-058-025.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21011 + }, + { + "path": "mallet_acoustic_047-091-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21012 + }, + { + "path": "mallet_acoustic_056-116-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_116", + "answer_gt": "midi_pitch_116", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21013 + }, + { + "path": "keyboard_electronic_069-032-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_55", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21014 + }, + { + "path": "reed_acoustic_037-062-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21015 + }, + { + "path": "organ_electronic_057-026-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21016 + }, + { + "path": "keyboard_acoustic_004-038-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21017 + }, + { + "path": "guitar_acoustic_030-045-050.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21018 + }, + { + "path": "guitar_acoustic_021-069-127.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21019 + }, + { + "path": "flute_synthetic_000-070-075.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21020 + }, + { + "path": "brass_acoustic_016-078-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_78", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_78", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21021 + }, + { + "path": "guitar_acoustic_010-058-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21022 + }, + { + "path": "guitar_acoustic_030-097-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_30", + "answer_gt": "midi_pitch_97", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21023 + }, + { + "path": "keyboard_acoustic_004-104-100.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_104", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21024 + }, + { + "path": "bass_synthetic_033-073-075.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_16", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21025 + }, + { + "path": "guitar_acoustic_015-067-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21026 + }, + { + "path": "keyboard_synthetic_000-081-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21027 + }, + { + "path": "keyboard_electronic_078-060-025.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21028 + }, + { + "path": "mallet_acoustic_062-061-050.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21029 + }, + { + "path": "flute_synthetic_000-095-025.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_90", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21030 + }, + { + "path": "brass_acoustic_059-040-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_40", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21031 + }, + { + "path": "mallet_acoustic_047-104-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21032 + }, + { + "path": "bass_synthetic_068-067-100.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21033 + }, + { + "path": "guitar_acoustic_010-054-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21034 + }, + { + "path": "organ_electronic_057-079-050.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21035 + }, + { + "path": "brass_acoustic_006-073-075.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21036 + }, + { + "path": "bass_synthetic_135-096-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21037 + }, + { + "path": "keyboard_synthetic_000-079-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21038 + }, + { + "path": "bass_synthetic_135-079-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21039 + }, + { + "path": "bass_electronic_027-032-127.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21040 + }, + { + "path": "keyboard_synthetic_000-021-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21041 + }, + { + "path": "guitar_electronic_028-087-075.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_35", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_87", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21042 + }, + { + "path": "organ_electronic_007-082-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21043 + }, + { + "path": "mallet_acoustic_062-062-025.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_9", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21044 + }, + { + "path": "keyboard_acoustic_004-050-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21045 + }, + { + "path": "keyboard_electronic_098-024-075.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_24", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21046 + }, + { + "path": "organ_electronic_007-048-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21047 + }, + { + "path": "vocal_synthetic_003-048-100.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_90", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21048 + }, + { + "path": "bass_synthetic_034-060-127.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21049 + }, + { + "path": "bass_synthetic_009-092-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21050 + }, + { + "path": "reed_acoustic_023-046-050.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21051 + }, + { + "path": "keyboard_electronic_098-099-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_99", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21052 + }, + { + "path": "bass_synthetic_034-064-100.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21053 + }, + { + "path": "mallet_acoustic_062-071-025.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21054 + }, + { + "path": "reed_acoustic_023-085-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21055 + }, + { + "path": "reed_acoustic_018-084-075.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21056 + }, + { + "path": "keyboard_electronic_069-053-100.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21057 + }, + { + "path": "brass_acoustic_016-088-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21058 + }, + { + "path": "bass_synthetic_098-065-050.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21059 + }, + { + "path": "guitar_electronic_028-094-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21060 + }, + { + "path": "brass_acoustic_046-081-100.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21061 + }, + { + "path": "bass_synthetic_135-096-025.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21062 + }, + { + "path": "keyboard_acoustic_004-089-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21063 + }, + { + "path": "string_acoustic_071-044-050.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21064 + }, + { + "path": "bass_synthetic_033-042-025.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21065 + }, + { + "path": "bass_synthetic_134-026-127.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21066 + }, + { + "path": "brass_acoustic_059-063-127.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21067 + }, + { + "path": "keyboard_synthetic_000-040-075.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21068 + }, + { + "path": "organ_electronic_007-077-127.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21069 + }, + { + "path": "bass_synthetic_135-070-025.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21070 + }, + { + "path": "bass_electronic_027-038-075.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21071 + }, + { + "path": "bass_electronic_025-036-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21072 + }, + { + "path": "guitar_acoustic_014-088-050.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21073 + }, + { + "path": "keyboard_acoustic_004-057-127.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21074 + }, + { + "path": "reed_acoustic_018-082-050.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21075 + }, + { + "path": "guitar_electronic_028-106-075.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21076 + }, + { + "path": "bass_synthetic_135-039-100.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21077 + }, + { + "path": "vocal_synthetic_003-067-100.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_90", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21078 + }, + { + "path": "brass_acoustic_006-054-127.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21079 + }, + { + "path": "keyboard_electronic_003-046-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_90", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21080 + }, + { + "path": "keyboard_electronic_001-064-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21081 + }, + { + "path": "bass_synthetic_033-077-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_35", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21082 + }, + { + "path": "mallet_acoustic_062-052-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21083 + }, + { + "path": "brass_acoustic_046-062-025.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_62", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21084 + }, + { + "path": "mallet_acoustic_062-038-025.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21085 + }, + { + "path": "keyboard_electronic_098-071-127.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21086 + }, + { + "path": "bass_electronic_018-060-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_20", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21087 + }, + { + "path": "keyboard_electronic_069-060-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21088 + }, + { + "path": "bass_synthetic_009-045-025.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_9", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21089 + }, + { + "path": "bass_electronic_025-060-025.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21090 + }, + { + "path": "keyboard_electronic_001-072-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21091 + }, + { + "path": "bass_synthetic_134-069-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21092 + }, + { + "path": "bass_synthetic_009-014-050.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_78", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_14", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21093 + }, + { + "path": "bass_synthetic_034-023-100.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21094 + }, + { + "path": "brass_acoustic_015-070-127.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_70", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21095 + }, + { + "path": "organ_electronic_104-068-025.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21096 + }, + { + "path": "bass_synthetic_068-037-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21097 + }, + { + "path": "bass_synthetic_098-050-100.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21098 + }, + { + "path": "vocal_synthetic_003-021-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21099 + }, + { + "path": "reed_acoustic_023-097-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_16", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_97", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21100 + }, + { + "path": "bass_synthetic_068-045-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21101 + }, + { + "path": "bass_synthetic_134-040-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21102 + }, + { + "path": "bass_synthetic_134-093-127.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21103 + }, + { + "path": "brass_acoustic_006-065-025.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21104 + }, + { + "path": "bass_synthetic_135-045-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21105 + }, + { + "path": "keyboard_electronic_098-086-050.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_24", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21106 + }, + { + "path": "mallet_acoustic_056-082-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_16", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21107 + }, + { + "path": "bass_synthetic_098-077-025.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21108 + }, + { + "path": "flute_synthetic_000-088-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21109 + }, + { + "path": "guitar_acoustic_010-051-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_30", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21110 + }, + { + "path": "brass_acoustic_046-071-025.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21111 + }, + { + "path": "organ_electronic_001-040-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21112 + }, + { + "path": "bass_synthetic_134-105-050.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_105", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21113 + }, + { + "path": "keyboard_electronic_001-022-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_22", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21114 + }, + { + "path": "reed_acoustic_011-053-100.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21115 + }, + { + "path": "bass_electronic_027-029-050.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21116 + }, + { + "path": "organ_electronic_007-042-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21117 + }, + { + "path": "keyboard_electronic_069-048-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21118 + }, + { + "path": "bass_synthetic_033-061-127.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21119 + }, + { + "path": "flute_synthetic_000-103-025.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21120 + }, + { + "path": "bass_synthetic_134-050-050.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21121 + }, + { + "path": "organ_electronic_007-077-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21122 + }, + { + "path": "bass_synthetic_009-044-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21123 + }, + { + "path": "guitar_acoustic_010-108-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_108", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21124 + }, + { + "path": "organ_electronic_028-053-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21125 + }, + { + "path": "guitar_electronic_022-059-050.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21126 + }, + { + "path": "guitar_electronic_028-077-075.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21127 + }, + { + "path": "string_acoustic_080-039-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_55", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21128 + }, + { + "path": "brass_acoustic_006-068-025.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21129 + }, + { + "path": "keyboard_electronic_098-038-127.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21130 + }, + { + "path": "guitar_electronic_028-044-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21131 + }, + { + "path": "keyboard_electronic_002-084-025.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21132 + }, + { + "path": "mallet_acoustic_062-049-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21133 + }, + { + "path": "mallet_acoustic_062-080-025.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21134 + }, + { + "path": "organ_electronic_113-075-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21135 + }, + { + "path": "mallet_acoustic_062-028-025.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21136 + }, + { + "path": "organ_electronic_104-032-050.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21137 + }, + { + "path": "mallet_acoustic_056-035-050.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21138 + }, + { + "path": "reed_acoustic_018-063-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21139 + }, + { + "path": "string_acoustic_014-041-100.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21140 + }, + { + "path": "string_acoustic_057-061-100.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21141 + }, + { + "path": "reed_acoustic_037-066-100.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21142 + }, + { + "path": "keyboard_acoustic_004-067-100.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21143 + }, + { + "path": "mallet_acoustic_062-058-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21144 + }, + { + "path": "brass_acoustic_016-068-100.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_74", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21145 + }, + { + "path": "keyboard_acoustic_004-023-025.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_104", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21146 + }, + { + "path": "organ_electronic_104-056-025.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21147 + }, + { + "path": "mallet_acoustic_056-058-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21148 + }, + { + "path": "organ_electronic_104-052-025.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21149 + }, + { + "path": "reed_acoustic_018-062-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21150 + }, + { + "path": "keyboard_electronic_098-056-050.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_102", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21151 + }, + { + "path": "mallet_acoustic_062-057-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21152 + }, + { + "path": "bass_synthetic_098-041-075.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21153 + }, + { + "path": "brass_acoustic_016-056-025.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21154 + }, + { + "path": "brass_acoustic_046-106-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21155 + }, + { + "path": "organ_electronic_007-081-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21156 + }, + { + "path": "flute_synthetic_000-040-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21157 + }, + { + "path": "bass_synthetic_068-106-025.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21158 + }, + { + "path": "keyboard_electronic_001-040-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21159 + }, + { + "path": "bass_synthetic_068-049-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21160 + }, + { + "path": "string_acoustic_056-050-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21161 + }, + { + "path": "bass_electronic_025-028-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21162 + }, + { + "path": "organ_electronic_028-098-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21163 + }, + { + "path": "string_acoustic_056-068-025.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21164 + }, + { + "path": "guitar_acoustic_021-076-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21165 + }, + { + "path": "flute_synthetic_000-072-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21166 + }, + { + "path": "keyboard_electronic_001-044-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21167 + }, + { + "path": "keyboard_electronic_098-065-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21168 + }, + { + "path": "bass_synthetic_033-057-100.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21169 + }, + { + "path": "guitar_acoustic_030-085-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21170 + }, + { + "path": "guitar_acoustic_021-091-127.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21171 + }, + { + "path": "bass_synthetic_009-012-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_12", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21172 + }, + { + "path": "organ_electronic_057-076-127.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21173 + }, + { + "path": "guitar_electronic_028-045-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21174 + }, + { + "path": "keyboard_electronic_098-079-127.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21175 + }, + { + "path": "keyboard_electronic_003-037-050.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_11", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21176 + }, + { + "path": "reed_acoustic_018-067-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21177 + }, + { + "path": "guitar_electronic_028-096-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21178 + }, + { + "path": "organ_electronic_007-051-127.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21179 + }, + { + "path": "keyboard_electronic_001-089-050.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21180 + }, + { + "path": "organ_electronic_001-064-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_30", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21181 + }, + { + "path": "guitar_acoustic_014-031-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21182 + }, + { + "path": "reed_acoustic_011-063-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21183 + }, + { + "path": "flute_synthetic_000-074-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21184 + }, + { + "path": "mallet_acoustic_062-045-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21185 + }, + { + "path": "bass_synthetic_134-108-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_108", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21186 + }, + { + "path": "keyboard_acoustic_004-040-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21187 + }, + { + "path": "reed_acoustic_023-073-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21188 + }, + { + "path": "bass_synthetic_068-032-050.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21189 + }, + { + "path": "keyboard_acoustic_004-046-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_10", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21190 + }, + { + "path": "keyboard_synthetic_000-032-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21191 + }, + { + "path": "flute_acoustic_002-073-127.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21192 + }, + { + "path": "bass_synthetic_098-049-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21193 + }, + { + "path": "keyboard_electronic_003-034-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21194 + }, + { + "path": "flute_synthetic_000-090-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_90", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21195 + }, + { + "path": "mallet_acoustic_062-058-100.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21196 + }, + { + "path": "reed_acoustic_037-073-025.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_20", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21197 + }, + { + "path": "string_acoustic_056-056-050.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21198 + }, + { + "path": "bass_synthetic_034-094-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21199 + }, + { + "path": "mallet_acoustic_047-104-025.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21200 + }, + { + "path": "string_acoustic_080-028-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21201 + }, + { + "path": "keyboard_electronic_098-107-100.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21202 + }, + { + "path": "brass_acoustic_015-077-050.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21203 + }, + { + "path": "bass_electronic_027-036-100.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21204 + }, + { + "path": "bass_synthetic_135-091-127.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21205 + }, + { + "path": "organ_electronic_028-088-075.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21206 + }, + { + "path": "organ_electronic_104-038-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21207 + }, + { + "path": "keyboard_electronic_003-072-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21208 + }, + { + "path": "vocal_synthetic_003-039-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21209 + }, + { + "path": "bass_synthetic_068-070-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21210 + }, + { + "path": "flute_synthetic_000-068-100.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21211 + }, + { + "path": "bass_synthetic_033-099-100.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_99", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21212 + }, + { + "path": "keyboard_acoustic_004-037-100.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21213 + }, + { + "path": "bass_synthetic_135-034-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21214 + }, + { + "path": "keyboard_acoustic_004-077-025.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21215 + }, + { + "path": "mallet_acoustic_047-092-025.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21216 + }, + { + "path": "bass_synthetic_033-060-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21217 + }, + { + "path": "guitar_acoustic_010-075-127.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21218 + }, + { + "path": "organ_electronic_057-060-127.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21219 + }, + { + "path": "bass_synthetic_134-055-100.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21220 + }, + { + "path": "organ_electronic_104-032-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21221 + }, + { + "path": "keyboard_acoustic_004-030-050.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_30", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21222 + }, + { + "path": "guitar_acoustic_014-106-050.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21223 + }, + { + "path": "bass_synthetic_135-073-127.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21224 + }, + { + "path": "mallet_acoustic_062-095-127.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_40", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21225 + }, + { + "path": "bass_synthetic_033-107-127.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21226 + }, + { + "path": "keyboard_electronic_069-045-100.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_19", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21227 + }, + { + "path": "bass_synthetic_009-096-127.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_116", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21228 + }, + { + "path": "organ_electronic_104-041-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21229 + }, + { + "path": "brass_acoustic_059-035-100.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21230 + }, + { + "path": "bass_synthetic_033-031-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21231 + }, + { + "path": "flute_synthetic_000-043-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21232 + }, + { + "path": "bass_synthetic_033-104-075.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21233 + }, + { + "path": "brass_acoustic_059-032-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21234 + }, + { + "path": "flute_synthetic_000-071-127.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21235 + }, + { + "path": "guitar_electronic_028-044-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21236 + }, + { + "path": "string_acoustic_056-039-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21237 + }, + { + "path": "reed_acoustic_037-075-050.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21238 + }, + { + "path": "bass_synthetic_033-090-025.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_90", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21239 + }, + { + "path": "keyboard_electronic_098-079-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21240 + }, + { + "path": "guitar_acoustic_030-041-100.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21241 + }, + { + "path": "reed_acoustic_023-043-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_25", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21242 + }, + { + "path": "organ_electronic_057-036-050.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_14", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21243 + }, + { + "path": "bass_synthetic_033-025-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_25", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21244 + }, + { + "path": "organ_electronic_007-044-100.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21245 + }, + { + "path": "string_acoustic_014-062-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21246 + }, + { + "path": "flute_acoustic_002-077-050.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21247 + }, + { + "path": "string_acoustic_014-029-100.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_16", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21248 + }, + { + "path": "reed_acoustic_023-071-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21249 + }, + { + "path": "keyboard_electronic_003-079-050.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21250 + }, + { + "path": "keyboard_electronic_001-065-127.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21251 + }, + { + "path": "bass_synthetic_033-089-050.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21252 + }, + { + "path": "reed_acoustic_023-039-050.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21253 + }, + { + "path": "keyboard_electronic_069-078-025.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_78", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21254 + }, + { + "path": "keyboard_electronic_098-028-075.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21255 + }, + { + "path": "keyboard_electronic_003-086-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21256 + }, + { + "path": "keyboard_acoustic_004-061-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21257 + }, + { + "path": "bass_synthetic_134-075-025.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21258 + }, + { + "path": "guitar_acoustic_014-022-025.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_22", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21259 + }, + { + "path": "organ_electronic_007-037-025.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21260 + }, + { + "path": "vocal_synthetic_003-091-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21261 + }, + { + "path": "reed_acoustic_023-056-050.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21262 + }, + { + "path": "guitar_acoustic_010-059-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21263 + }, + { + "path": "flute_synthetic_000-061-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21264 + }, + { + "path": "organ_electronic_104-042-050.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21265 + }, + { + "path": "flute_acoustic_002-074-127.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21266 + }, + { + "path": "guitar_electronic_028-101-127.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21267 + }, + { + "path": "bass_synthetic_033-049-127.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21268 + }, + { + "path": "bass_synthetic_135-050-025.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21269 + }, + { + "path": "brass_acoustic_016-080-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21270 + }, + { + "path": "string_acoustic_056-075-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21271 + }, + { + "path": "guitar_electronic_028-057-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21272 + }, + { + "path": "string_acoustic_056-049-025.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21273 + }, + { + "path": "bass_synthetic_068-027-127.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21274 + }, + { + "path": "string_acoustic_057-063-025.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21275 + }, + { + "path": "guitar_acoustic_010-107-075.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21276 + }, + { + "path": "mallet_acoustic_056-083-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21277 + }, + { + "path": "organ_electronic_028-040-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21278 + }, + { + "path": "guitar_acoustic_014-107-025.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_16", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21279 + }, + { + "path": "guitar_electronic_022-075-127.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21280 + }, + { + "path": "bass_synthetic_034-021-025.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21281 + }, + { + "path": "bass_electronic_027-046-025.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_90", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21282 + }, + { + "path": "flute_synthetic_000-030-075.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21283 + }, + { + "path": "bass_synthetic_098-073-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21284 + }, + { + "path": "guitar_acoustic_030-023-025.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21285 + }, + { + "path": "bass_synthetic_033-061-025.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21286 + }, + { + "path": "flute_synthetic_000-104-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_20", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21287 + }, + { + "path": "organ_electronic_028-050-100.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21288 + }, + { + "path": "keyboard_electronic_002-103-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21289 + }, + { + "path": "guitar_acoustic_021-056-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_104", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21290 + }, + { + "path": "string_acoustic_014-048-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21291 + }, + { + "path": "guitar_acoustic_015-075-025.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21292 + }, + { + "path": "organ_electronic_028-106-100.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21293 + }, + { + "path": "bass_synthetic_134-032-100.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21294 + }, + { + "path": "brass_acoustic_016-057-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21295 + }, + { + "path": "keyboard_electronic_001-073-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21296 + }, + { + "path": "reed_acoustic_023-068-025.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21297 + }, + { + "path": "bass_synthetic_009-060-075.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21298 + }, + { + "path": "keyboard_electronic_098-066-025.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21299 + }, + { + "path": "organ_electronic_001-095-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21300 + }, + { + "path": "bass_electronic_018-043-050.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21301 + }, + { + "path": "string_acoustic_080-052-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_76", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21302 + }, + { + "path": "bass_synthetic_068-038-127.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21303 + }, + { + "path": "guitar_acoustic_015-057-025.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21304 + }, + { + "path": "string_acoustic_014-051-025.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21305 + }, + { + "path": "keyboard_electronic_098-063-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21306 + }, + { + "path": "brass_acoustic_006-082-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21307 + }, + { + "path": "organ_electronic_028-031-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21308 + }, + { + "path": "keyboard_electronic_078-041-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21309 + }, + { + "path": "guitar_electronic_028-067-127.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21310 + }, + { + "path": "keyboard_electronic_098-043-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21311 + }, + { + "path": "organ_electronic_007-066-075.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_20", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21312 + }, + { + "path": "keyboard_synthetic_000-093-127.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21313 + }, + { + "path": "guitar_acoustic_014-105-075.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_105", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21314 + }, + { + "path": "guitar_electronic_028-097-127.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_97", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21315 + }, + { + "path": "organ_electronic_001-077-075.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21316 + }, + { + "path": "organ_electronic_104-083-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21317 + }, + { + "path": "guitar_electronic_028-060-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21318 + }, + { + "path": "keyboard_electronic_078-038-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21319 + }, + { + "path": "bass_electronic_018-052-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21320 + }, + { + "path": "guitar_electronic_028-095-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_76", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21321 + }, + { + "path": "vocal_synthetic_003-044-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21322 + }, + { + "path": "brass_acoustic_015-046-025.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21323 + }, + { + "path": "reed_acoustic_037-058-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21324 + }, + { + "path": "bass_synthetic_009-070-100.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21325 + }, + { + "path": "reed_acoustic_023-076-127.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21326 + }, + { + "path": "string_acoustic_057-088-075.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21327 + }, + { + "path": "bass_electronic_025-063-075.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21328 + }, + { + "path": "keyboard_electronic_003-023-025.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21329 + }, + { + "path": "bass_synthetic_098-066-025.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21330 + }, + { + "path": "keyboard_electronic_001-042-025.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21331 + }, + { + "path": "brass_acoustic_016-081-127.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21332 + }, + { + "path": "mallet_acoustic_062-075-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21333 + }, + { + "path": "guitar_acoustic_014-053-100.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21334 + }, + { + "path": "bass_synthetic_009-077-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_70", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21335 + }, + { + "path": "keyboard_acoustic_004-041-050.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21336 + }, + { + "path": "bass_synthetic_009-082-050.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21337 + }, + { + "path": "bass_synthetic_098-041-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21338 + }, + { + "path": "string_acoustic_014-065-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21339 + }, + { + "path": "organ_electronic_104-050-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21340 + }, + { + "path": "string_acoustic_071-035-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21341 + }, + { + "path": "guitar_acoustic_010-093-050.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21342 + }, + { + "path": "guitar_acoustic_015-077-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21343 + }, + { + "path": "guitar_acoustic_015-071-127.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_102", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21344 + }, + { + "path": "bass_electronic_025-028-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21345 + }, + { + "path": "vocal_acoustic_000-054-050.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21346 + }, + { + "path": "organ_electronic_007-086-050.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21347 + }, + { + "path": "organ_electronic_113-045-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21348 + }, + { + "path": "reed_acoustic_018-065-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21349 + }, + { + "path": "keyboard_synthetic_000-026-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21350 + }, + { + "path": "mallet_acoustic_062-024-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_24", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21351 + }, + { + "path": "flute_synthetic_000-036-100.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21352 + }, + { + "path": "string_acoustic_071-029-025.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21353 + }, + { + "path": "bass_synthetic_135-046-127.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21354 + }, + { + "path": "bass_synthetic_009-067-050.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21355 + }, + { + "path": "reed_acoustic_011-035-025.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21356 + }, + { + "path": "guitar_acoustic_015-054-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21357 + }, + { + "path": "mallet_acoustic_062-079-025.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21358 + }, + { + "path": "guitar_acoustic_014-052-127.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21359 + }, + { + "path": "vocal_synthetic_003-054-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21360 + }, + { + "path": "organ_electronic_001-069-025.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21361 + }, + { + "path": "string_acoustic_056-040-075.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_10", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21362 + }, + { + "path": "bass_synthetic_068-069-100.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21363 + }, + { + "path": "organ_electronic_104-033-100.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21364 + }, + { + "path": "keyboard_electronic_098-044-025.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_11", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21365 + }, + { + "path": "organ_electronic_113-070-025.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21366 + }, + { + "path": "bass_synthetic_135-027-025.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21367 + }, + { + "path": "reed_acoustic_011-045-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21368 + }, + { + "path": "mallet_acoustic_062-039-075.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21369 + }, + { + "path": "bass_electronic_025-032-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21370 + }, + { + "path": "guitar_acoustic_014-102-025.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_102", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21371 + }, + { + "path": "keyboard_electronic_003-092-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21372 + }, + { + "path": "bass_synthetic_098-059-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21373 + }, + { + "path": "bass_synthetic_033-065-050.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21374 + }, + { + "path": "guitar_acoustic_014-082-075.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21375 + }, + { + "path": "guitar_electronic_022-049-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21376 + }, + { + "path": "guitar_acoustic_015-044-100.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21377 + }, + { + "path": "bass_synthetic_009-041-075.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_55", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21378 + }, + { + "path": "keyboard_synthetic_000-025-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21379 + }, + { + "path": "string_acoustic_056-070-050.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21380 + }, + { + "path": "string_acoustic_014-045-100.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21381 + }, + { + "path": "keyboard_electronic_001-032-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21382 + }, + { + "path": "keyboard_electronic_069-076-075.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21383 + }, + { + "path": "organ_electronic_001-087-075.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_87", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21384 + }, + { + "path": "flute_acoustic_002-103-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21385 + }, + { + "path": "keyboard_electronic_078-045-025.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21386 + }, + { + "path": "brass_acoustic_046-057-050.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21387 + }, + { + "path": "organ_electronic_104-082-100.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21388 + }, + { + "path": "guitar_acoustic_015-074-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_74", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_116", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21389 + }, + { + "path": "organ_electronic_007-011-075.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_11", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21390 + }, + { + "path": "keyboard_electronic_003-073-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21391 + }, + { + "path": "keyboard_electronic_002-076-075.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21392 + }, + { + "path": "guitar_acoustic_015-098-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21393 + }, + { + "path": "brass_acoustic_016-085-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21394 + }, + { + "path": "reed_acoustic_023-033-075.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21395 + }, + { + "path": "bass_synthetic_009-081-100.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_14", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21396 + }, + { + "path": "keyboard_electronic_069-061-075.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21397 + }, + { + "path": "mallet_acoustic_062-104-050.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_10", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21398 + }, + { + "path": "keyboard_electronic_003-089-075.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21399 + }, + { + "path": "mallet_acoustic_047-065-050.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_25", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21400 + }, + { + "path": "string_acoustic_080-053-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21401 + }, + { + "path": "bass_synthetic_134-085-050.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21402 + }, + { + "path": "reed_acoustic_023-052-025.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21403 + }, + { + "path": "brass_acoustic_016-067-075.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21404 + }, + { + "path": "mallet_acoustic_056-044-025.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21405 + }, + { + "path": "vocal_synthetic_003-049-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21406 + }, + { + "path": "organ_electronic_104-075-025.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21407 + }, + { + "path": "guitar_acoustic_021-035-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21408 + }, + { + "path": "guitar_electronic_022-042-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21409 + }, + { + "path": "guitar_acoustic_015-058-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21410 + }, + { + "path": "guitar_electronic_022-078-050.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_78", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21411 + }, + { + "path": "organ_electronic_113-083-075.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21412 + }, + { + "path": "bass_synthetic_098-090-050.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_90", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21413 + }, + { + "path": "string_acoustic_071-038-025.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21414 + }, + { + "path": "keyboard_acoustic_004-084-025.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21415 + }, + { + "path": "keyboard_electronic_001-092-050.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_102", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21416 + }, + { + "path": "guitar_electronic_028-029-075.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21417 + }, + { + "path": "guitar_acoustic_021-025-127.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21418 + }, + { + "path": "keyboard_synthetic_000-064-127.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21419 + }, + { + "path": "keyboard_electronic_003-038-075.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21420 + }, + { + "path": "bass_electronic_025-038-025.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_91", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21421 + }, + { + "path": "keyboard_electronic_002-059-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21422 + }, + { + "path": "bass_synthetic_068-077-127.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21423 + }, + { + "path": "bass_synthetic_068-029-075.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_35", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21424 + }, + { + "path": "vocal_synthetic_003-027-025.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21425 + }, + { + "path": "guitar_electronic_028-030-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21426 + }, + { + "path": "string_acoustic_057-067-100.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21427 + }, + { + "path": "keyboard_electronic_003-025-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21428 + }, + { + "path": "reed_acoustic_037-068-050.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_23", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21429 + }, + { + "path": "keyboard_electronic_001-043-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21430 + }, + { + "path": "guitar_acoustic_014-069-127.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21431 + }, + { + "path": "keyboard_acoustic_004-096-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21432 + }, + { + "path": "mallet_acoustic_062-083-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21433 + }, + { + "path": "guitar_acoustic_030-077-100.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21434 + }, + { + "path": "organ_electronic_001-061-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21435 + }, + { + "path": "bass_synthetic_134-036-075.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21436 + }, + { + "path": "mallet_acoustic_062-098-025.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21437 + }, + { + "path": "guitar_acoustic_010-070-025.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21438 + }, + { + "path": "flute_synthetic_000-105-100.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_105", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21439 + }, + { + "path": "organ_electronic_028-066-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21440 + }, + { + "path": "keyboard_electronic_003-056-127.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21441 + }, + { + "path": "keyboard_electronic_001-060-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21442 + }, + { + "path": "brass_acoustic_006-068-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_113", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21443 + }, + { + "path": "string_acoustic_012-026-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21444 + }, + { + "path": "brass_acoustic_006-073-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21445 + }, + { + "path": "flute_synthetic_000-077-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21446 + }, + { + "path": "vocal_synthetic_003-101-100.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_62", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21447 + }, + { + "path": "bass_synthetic_134-052-050.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_30", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21448 + }, + { + "path": "organ_electronic_104-045-075.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21449 + }, + { + "path": "keyboard_electronic_003-061-075.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21450 + }, + { + "path": "bass_synthetic_134-045-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_17", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21451 + }, + { + "path": "keyboard_electronic_001-052-025.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21452 + }, + { + "path": "vocal_synthetic_003-106-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21453 + }, + { + "path": "keyboard_electronic_098-089-025.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_13", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21454 + }, + { + "path": "bass_synthetic_033-101-050.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21455 + }, + { + "path": "bass_synthetic_134-074-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21456 + }, + { + "path": "organ_electronic_028-031-100.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21457 + }, + { + "path": "guitar_acoustic_010-042-050.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_13", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21458 + }, + { + "path": "brass_acoustic_016-066-075.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21459 + }, + { + "path": "organ_electronic_113-051-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21460 + }, + { + "path": "mallet_acoustic_062-057-025.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21461 + }, + { + "path": "bass_synthetic_098-028-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21462 + }, + { + "path": "mallet_acoustic_056-113-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_113", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21463 + }, + { + "path": "brass_acoustic_059-062-025.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21464 + }, + { + "path": "keyboard_synthetic_000-070-127.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_70", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21465 + }, + { + "path": "keyboard_electronic_098-098-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21466 + }, + { + "path": "string_acoustic_014-034-075.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21467 + }, + { + "path": "organ_electronic_028-066-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21468 + }, + { + "path": "guitar_acoustic_010-074-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_74", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21469 + }, + { + "path": "guitar_electronic_028-032-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21470 + }, + { + "path": "guitar_acoustic_010-030-127.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21471 + }, + { + "path": "organ_electronic_113-075-025.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21472 + }, + { + "path": "organ_electronic_007-043-075.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21473 + }, + { + "path": "guitar_acoustic_030-107-050.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_116", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_107", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21474 + }, + { + "path": "organ_electronic_104-047-025.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21475 + }, + { + "path": "string_acoustic_056-059-075.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21476 + }, + { + "path": "guitar_acoustic_021-069-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21477 + }, + { + "path": "organ_electronic_007-076-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_76", + "choice_d": "midi_pitch_23", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21478 + }, + { + "path": "bass_synthetic_134-075-050.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21479 + }, + { + "path": "mallet_acoustic_047-105-075.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_105", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21480 + }, + { + "path": "keyboard_electronic_001-054-025.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21481 + }, + { + "path": "bass_synthetic_068-075-025.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21482 + }, + { + "path": "keyboard_electronic_001-086-075.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_28", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21483 + }, + { + "path": "keyboard_electronic_003-047-127.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21484 + }, + { + "path": "keyboard_electronic_069-066-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21485 + }, + { + "path": "flute_acoustic_002-070-025.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_14", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21486 + }, + { + "path": "bass_synthetic_034-050-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21487 + }, + { + "path": "guitar_acoustic_014-086-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_104", + "choice_d": "midi_pitch_44", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21488 + }, + { + "path": "keyboard_electronic_002-063-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21489 + }, + { + "path": "keyboard_electronic_078-046-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_106", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21490 + }, + { + "path": "bass_electronic_018-023-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21491 + }, + { + "path": "mallet_acoustic_056-075-050.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21492 + }, + { + "path": "keyboard_electronic_003-028-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21493 + }, + { + "path": "keyboard_electronic_069-059-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21494 + }, + { + "path": "keyboard_electronic_001-104-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_104", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21495 + }, + { + "path": "keyboard_synthetic_000-094-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21496 + }, + { + "path": "organ_electronic_113-071-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_11", + "choice_b": "midi_pitch_20", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21497 + }, + { + "path": "guitar_acoustic_021-095-075.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21498 + }, + { + "path": "guitar_acoustic_015-067-050.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21499 + }, + { + "path": "keyboard_electronic_001-031-050.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21500 + }, + { + "path": "bass_synthetic_098-070-050.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21501 + }, + { + "path": "guitar_acoustic_015-059-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21502 + }, + { + "path": "bass_synthetic_034-023-075.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21503 + }, + { + "path": "bass_synthetic_135-049-050.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21504 + }, + { + "path": "keyboard_synthetic_000-024-075.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_24", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21505 + }, + { + "path": "string_acoustic_056-045-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21506 + }, + { + "path": "guitar_acoustic_010-035-025.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_90", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21507 + }, + { + "path": "guitar_acoustic_015-069-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21508 + }, + { + "path": "brass_acoustic_015-076-025.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21509 + }, + { + "path": "guitar_electronic_022-047-100.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_14", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21510 + }, + { + "path": "reed_acoustic_023-065-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21511 + }, + { + "path": "brass_acoustic_006-027-050.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21512 + }, + { + "path": "flute_acoustic_002-084-050.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21513 + }, + { + "path": "brass_acoustic_006-026-127.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21514 + }, + { + "path": "guitar_acoustic_015-030-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21515 + }, + { + "path": "keyboard_electronic_069-064-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21516 + }, + { + "path": "bass_synthetic_134-080-075.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_62", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21517 + }, + { + "path": "guitar_acoustic_015-104-100.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_104", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21518 + }, + { + "path": "organ_electronic_113-075-100.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21519 + }, + { + "path": "guitar_acoustic_014-108-050.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_108", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21520 + }, + { + "path": "organ_electronic_007-041-100.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21521 + }, + { + "path": "reed_acoustic_018-079-075.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_19", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21522 + }, + { + "path": "keyboard_synthetic_000-076-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21523 + }, + { + "path": "keyboard_electronic_078-083-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21524 + }, + { + "path": "bass_synthetic_009-080-100.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21525 + }, + { + "path": "keyboard_electronic_098-049-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21526 + }, + { + "path": "guitar_acoustic_014-072-100.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21527 + }, + { + "path": "keyboard_electronic_098-042-025.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21528 + }, + { + "path": "guitar_acoustic_015-048-100.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21529 + }, + { + "path": "bass_synthetic_134-081-025.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21530 + }, + { + "path": "vocal_synthetic_003-084-025.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21531 + }, + { + "path": "organ_electronic_007-070-075.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_58", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21532 + }, + { + "path": "reed_acoustic_018-061-127.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21533 + }, + { + "path": "organ_electronic_007-080-050.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21534 + }, + { + "path": "keyboard_electronic_002-083-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_83", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21535 + }, + { + "path": "keyboard_electronic_001-059-075.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21536 + }, + { + "path": "keyboard_electronic_003-082-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_105", + "choice_b": "midi_pitch_97", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21537 + }, + { + "path": "guitar_acoustic_014-039-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21538 + }, + { + "path": "reed_acoustic_018-083-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21539 + }, + { + "path": "vocal_synthetic_003-061-025.wav", + "question": "Identify the MIDI pitch of the single musical note.", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21540 + }, + { + "path": "bass_electronic_025-035-127.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_35", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21541 + }, + { + "path": "brass_acoustic_059-031-050.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21542 + }, + { + "path": "keyboard_synthetic_000-040-050.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21543 + }, + { + "path": "keyboard_electronic_003-023-100.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21544 + }, + { + "path": "keyboard_acoustic_004-091-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21545 + }, + { + "path": "string_acoustic_057-056-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21546 + }, + { + "path": "brass_acoustic_046-095-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_76", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21547 + }, + { + "path": "reed_acoustic_037-035-050.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_35", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21548 + }, + { + "path": "bass_synthetic_033-085-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21549 + }, + { + "path": "guitar_acoustic_014-027-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21550 + }, + { + "path": "bass_synthetic_135-055-127.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21551 + }, + { + "path": "organ_electronic_104-067-127.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_67", + "choice_c": "midi_pitch_19", + "choice_d": "midi_pitch_65", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21552 + }, + { + "path": "organ_electronic_007-092-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21553 + }, + { + "path": "vocal_synthetic_003-039-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21554 + }, + { + "path": "brass_acoustic_016-061-100.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21555 + }, + { + "path": "flute_acoustic_002-095-100.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21556 + }, + { + "path": "string_acoustic_071-049-025.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21557 + }, + { + "path": "guitar_acoustic_030-079-127.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21558 + }, + { + "path": "keyboard_electronic_069-084-100.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21559 + }, + { + "path": "bass_synthetic_068-046-075.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21560 + }, + { + "path": "keyboard_synthetic_000-062-100.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21561 + }, + { + "path": "reed_acoustic_011-067-100.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21562 + }, + { + "path": "keyboard_synthetic_000-034-050.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21563 + }, + { + "path": "organ_electronic_104-039-025.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21564 + }, + { + "path": "bass_synthetic_009-013-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_13", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21565 + }, + { + "path": "keyboard_electronic_098-043-075.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21566 + }, + { + "path": "guitar_acoustic_015-094-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21567 + }, + { + "path": "bass_synthetic_033-064-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_19", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21568 + }, + { + "path": "reed_acoustic_023-086-075.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21569 + }, + { + "path": "reed_acoustic_011-068-025.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_13", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21570 + }, + { + "path": "guitar_acoustic_014-070-025.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_48", + "choice_d": "midi_pitch_78", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21571 + }, + { + "path": "keyboard_electronic_003-087-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_87", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21572 + }, + { + "path": "mallet_acoustic_062-084-100.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_16", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21573 + }, + { + "path": "organ_electronic_007-028-025.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_20", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21574 + }, + { + "path": "brass_acoustic_046-071-100.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21575 + }, + { + "path": "keyboard_electronic_069-048-075.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21576 + }, + { + "path": "keyboard_electronic_069-082-100.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21577 + }, + { + "path": "guitar_acoustic_010-057-100.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21578 + }, + { + "path": "keyboard_electronic_078-069-100.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21579 + }, + { + "path": "bass_synthetic_033-030-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21580 + }, + { + "path": "mallet_acoustic_047-103-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21581 + }, + { + "path": "organ_electronic_028-097-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_97", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21582 + }, + { + "path": "guitar_acoustic_010-093-100.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21583 + }, + { + "path": "string_acoustic_057-076-025.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21584 + }, + { + "path": "keyboard_electronic_002-024-050.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21585 + }, + { + "path": "string_acoustic_056-066-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21586 + }, + { + "path": "bass_synthetic_033-072-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_22", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21587 + }, + { + "path": "bass_synthetic_033-069-100.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_69", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21588 + }, + { + "path": "guitar_acoustic_021-040-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_40", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21589 + }, + { + "path": "bass_synthetic_009-036-025.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21590 + }, + { + "path": "keyboard_electronic_002-041-100.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_76", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21591 + }, + { + "path": "organ_electronic_057-034-100.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21592 + }, + { + "path": "bass_synthetic_009-052-100.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21593 + }, + { + "path": "bass_synthetic_134-098-100.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_98", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21594 + }, + { + "path": "organ_electronic_057-027-075.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21595 + }, + { + "path": "guitar_acoustic_010-039-127.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21596 + }, + { + "path": "flute_acoustic_002-092-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_92", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21597 + }, + { + "path": "string_acoustic_014-028-127.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_73", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21598 + }, + { + "path": "keyboard_electronic_069-050-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_54", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21599 + }, + { + "path": "keyboard_electronic_001-065-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_25", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21600 + }, + { + "path": "guitar_electronic_022-036-100.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21601 + }, + { + "path": "guitar_acoustic_014-100-075.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_100", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21602 + }, + { + "path": "bass_synthetic_098-061-050.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_106", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_61", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21603 + }, + { + "path": "guitar_electronic_022-038-050.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21604 + }, + { + "path": "bass_synthetic_033-099-127.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_99", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21605 + }, + { + "path": "keyboard_electronic_078-060-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21606 + }, + { + "path": "guitar_acoustic_015-082-075.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21607 + }, + { + "path": "guitar_acoustic_014-101-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21608 + }, + { + "path": "bass_electronic_018-027-075.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_27", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21609 + }, + { + "path": "bass_electronic_018-061-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_27", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21610 + }, + { + "path": "organ_electronic_113-053-075.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21611 + }, + { + "path": "bass_synthetic_135-068-075.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21612 + }, + { + "path": "string_acoustic_056-038-050.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21613 + }, + { + "path": "bass_synthetic_134-085-127.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21614 + }, + { + "path": "keyboard_electronic_001-057-127.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21615 + }, + { + "path": "brass_acoustic_006-077-100.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21616 + }, + { + "path": "string_acoustic_014-031-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_38", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21617 + }, + { + "path": "guitar_acoustic_015-103-127.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21618 + }, + { + "path": "reed_acoustic_037-041-075.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21619 + }, + { + "path": "bass_synthetic_068-098-025.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_14", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21620 + }, + { + "path": "bass_electronic_018-060-127.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21621 + }, + { + "path": "keyboard_electronic_003-053-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21622 + }, + { + "path": "organ_electronic_104-055-075.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_55", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21623 + }, + { + "path": "bass_synthetic_134-100-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_100", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21624 + }, + { + "path": "mallet_acoustic_062-050-075.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21625 + }, + { + "path": "keyboard_electronic_002-100-127.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_100", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21626 + }, + { + "path": "bass_synthetic_098-071-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_91", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21627 + }, + { + "path": "guitar_acoustic_010-084-025.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_26", + "choice_c": "midi_pitch_28", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21628 + }, + { + "path": "string_acoustic_057-069-075.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_116", + "answer_gt": "midi_pitch_69", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21629 + }, + { + "path": "organ_electronic_104-026-127.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21630 + }, + { + "path": "brass_acoustic_006-034-075.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21631 + }, + { + "path": "bass_synthetic_009-058-025.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21632 + }, + { + "path": "guitar_acoustic_014-096-100.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_96", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21633 + }, + { + "path": "organ_electronic_104-077-025.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_63", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21634 + }, + { + "path": "string_acoustic_057-059-025.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_59", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21635 + }, + { + "path": "bass_synthetic_034-093-075.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21636 + }, + { + "path": "string_acoustic_014-063-127.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21637 + }, + { + "path": "keyboard_electronic_078-061-127.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21638 + }, + { + "path": "bass_synthetic_034-085-075.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21639 + }, + { + "path": "reed_acoustic_018-067-025.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_67", + "choice_b": "midi_pitch_69", + "choice_c": "midi_pitch_102", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21640 + }, + { + "path": "bass_electronic_018-057-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21641 + }, + { + "path": "bass_synthetic_098-038-100.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_38", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21642 + }, + { + "path": "keyboard_electronic_003-071-100.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_15", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_47", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21643 + }, + { + "path": "organ_electronic_113-053-050.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_53", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21644 + }, + { + "path": "keyboard_electronic_002-089-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_87", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_89", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21645 + }, + { + "path": "organ_electronic_104-053-025.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21646 + }, + { + "path": "keyboard_acoustic_004-079-075.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21647 + }, + { + "path": "organ_electronic_028-042-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_24", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21648 + }, + { + "path": "guitar_acoustic_015-027-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21649 + }, + { + "path": "guitar_electronic_022-086-127.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_86", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21650 + }, + { + "path": "organ_electronic_007-096-075.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_96", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21651 + }, + { + "path": "keyboard_electronic_003-086-127.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21652 + }, + { + "path": "mallet_acoustic_047-095-127.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21653 + }, + { + "path": "bass_electronic_018-036-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21654 + }, + { + "path": "guitar_acoustic_010-021-127.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_89", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21655 + }, + { + "path": "reed_acoustic_011-074-127.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21656 + }, + { + "path": "guitar_electronic_028-079-127.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_79", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21657 + }, + { + "path": "bass_synthetic_034-085-025.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21658 + }, + { + "path": "organ_electronic_007-010-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_10", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21659 + }, + { + "path": "string_acoustic_012-055-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21660 + }, + { + "path": "guitar_acoustic_010-032-075.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21661 + }, + { + "path": "guitar_acoustic_021-090-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_49", + "choice_d": "midi_pitch_90", + "answer_gt": "midi_pitch_90", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21662 + }, + { + "path": "bass_synthetic_033-056-050.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_95", + "choice_c": "midi_pitch_56", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21663 + }, + { + "path": "guitar_electronic_022-073-127.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_73", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_51", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21664 + }, + { + "path": "guitar_acoustic_010-070-075.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_68", + "choice_c": "midi_pitch_70", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21665 + }, + { + "path": "keyboard_acoustic_004-031-075.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21666 + }, + { + "path": "brass_acoustic_006-056-127.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21667 + }, + { + "path": "keyboard_electronic_098-074-127.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_74", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21668 + }, + { + "path": "guitar_acoustic_010-085-100.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_14", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21669 + }, + { + "path": "string_acoustic_012-050-100.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_50", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21670 + }, + { + "path": "keyboard_electronic_002-025-025.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21671 + }, + { + "path": "brass_acoustic_046-081-050.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_81", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21672 + }, + { + "path": "keyboard_electronic_001-045-025.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21673 + }, + { + "path": "keyboard_acoustic_004-091-100.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_50", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21674 + }, + { + "path": "guitar_acoustic_015-021-075.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_53", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21675 + }, + { + "path": "keyboard_electronic_098-103-127.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21676 + }, + { + "path": "bass_synthetic_034-063-025.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_100", + "choice_b": "midi_pitch_63", + "choice_c": "midi_pitch_65", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21677 + }, + { + "path": "bass_synthetic_134-034-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21678 + }, + { + "path": "brass_acoustic_059-052-050.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_25", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21679 + }, + { + "path": "brass_acoustic_015-066-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_11", + "choice_d": "midi_pitch_66", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21680 + }, + { + "path": "bass_synthetic_098-061-127.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21681 + }, + { + "path": "organ_electronic_028-034-100.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_80", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21682 + }, + { + "path": "keyboard_electronic_001-077-075.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21683 + }, + { + "path": "brass_acoustic_059-048-050.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21684 + }, + { + "path": "keyboard_electronic_002-090-050.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_90", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_90", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21685 + }, + { + "path": "string_acoustic_014-050-075.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21686 + }, + { + "path": "bass_synthetic_135-087-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_87", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21687 + }, + { + "path": "bass_synthetic_009-019-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_31", + "answer_gt": "midi_pitch_19", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21688 + }, + { + "path": "keyboard_synthetic_000-098-075.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_99", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21689 + }, + { + "path": "mallet_acoustic_047-066-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_66", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21690 + }, + { + "path": "keyboard_electronic_069-039-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21691 + }, + { + "path": "vocal_synthetic_003-065-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21692 + }, + { + "path": "brass_acoustic_015-058-050.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_82", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21693 + }, + { + "path": "reed_acoustic_023-041-127.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_89", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21694 + }, + { + "path": "flute_synthetic_000-037-127.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21695 + }, + { + "path": "guitar_acoustic_010-024-050.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_24", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21696 + }, + { + "path": "string_acoustic_056-044-025.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21697 + }, + { + "path": "bass_electronic_018-023-050.wav", + "question": "Select the MIDI pitch that aligns with this note's sound.", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21698 + }, + { + "path": "keyboard_electronic_002-067-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_67", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21699 + }, + { + "path": "guitar_electronic_028-073-127.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21700 + }, + { + "path": "bass_synthetic_033-041-050.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21701 + }, + { + "path": "guitar_acoustic_021-092-127.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_48", + "choice_c": "midi_pitch_74", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_92", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21702 + }, + { + "path": "vocal_synthetic_003-037-075.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21703 + }, + { + "path": "reed_acoustic_011-043-075.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_101", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_54", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21704 + }, + { + "path": "guitar_electronic_028-093-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21705 + }, + { + "path": "organ_electronic_028-084-075.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21706 + }, + { + "path": "guitar_acoustic_021-046-127.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21707 + }, + { + "path": "keyboard_acoustic_004-061-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_13", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21708 + }, + { + "path": "keyboard_electronic_069-075-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_57", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21709 + }, + { + "path": "keyboard_acoustic_004-075-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_102", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_96", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21710 + }, + { + "path": "keyboard_electronic_098-044-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_95", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21711 + }, + { + "path": "keyboard_electronic_002-046-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21712 + }, + { + "path": "guitar_acoustic_021-048-025.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_48", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21713 + }, + { + "path": "organ_electronic_001-084-050.wav", + "question": "What is the MIDI pitch of the note in this music?", + "choice_a": "midi_pitch_30", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21714 + }, + { + "path": "vocal_synthetic_003-054-075.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21715 + }, + { + "path": "keyboard_electronic_098-045-075.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21716 + }, + { + "path": "organ_electronic_104-029-025.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21717 + }, + { + "path": "organ_electronic_001-076-050.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_40", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21718 + }, + { + "path": "keyboard_synthetic_000-033-025.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_20", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21719 + }, + { + "path": "brass_acoustic_006-057-100.wav", + "question": "The MIDI pitch of the played note is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21720 + }, + { + "path": "keyboard_acoustic_004-065-025.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_58", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21721 + }, + { + "path": "bass_synthetic_135-094-100.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21722 + }, + { + "path": "guitar_acoustic_015-108-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_108", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_108", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21723 + }, + { + "path": "bass_synthetic_134-075-075.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_52", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21724 + }, + { + "path": "reed_acoustic_011-054-025.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_39", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21725 + }, + { + "path": "brass_acoustic_015-054-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21726 + }, + { + "path": "brass_acoustic_046-104-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_55", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21727 + }, + { + "path": "reed_acoustic_018-084-127.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_84", + "choice_c": "midi_pitch_116", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21728 + }, + { + "path": "organ_electronic_007-082-100.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21729 + }, + { + "path": "bass_synthetic_033-045-075.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_37", + "choice_c": "midi_pitch_45", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21730 + }, + { + "path": "guitar_acoustic_030-064-025.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_64", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21731 + }, + { + "path": "bass_synthetic_033-038-127.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_38", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21732 + }, + { + "path": "bass_electronic_018-049-075.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_49", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_116", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21733 + }, + { + "path": "keyboard_synthetic_000-099-050.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_41", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_86", + "answer_gt": "midi_pitch_99", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21734 + }, + { + "path": "string_acoustic_014-026-127.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_13", + "choice_c": "midi_pitch_78", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21735 + }, + { + "path": "organ_electronic_007-080-100.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_17", + "choice_b": "midi_pitch_74", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21736 + }, + { + "path": "bass_synthetic_135-083-127.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_68", + "answer_gt": "midi_pitch_83", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21737 + }, + { + "path": "mallet_acoustic_047-068-050.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_64", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21738 + }, + { + "path": "string_acoustic_057-062-075.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21739 + }, + { + "path": "string_acoustic_057-081-025.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_79", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21740 + }, + { + "path": "bass_synthetic_009-047-100.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21741 + }, + { + "path": "bass_synthetic_009-094-100.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21742 + }, + { + "path": "organ_electronic_057-052-100.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21743 + }, + { + "path": "guitar_electronic_022-085-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_39", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21744 + }, + { + "path": "keyboard_electronic_002-101-075.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_101", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21745 + }, + { + "path": "bass_synthetic_068-046-025.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_46", + "choice_c": "midi_pitch_32", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21746 + }, + { + "path": "keyboard_electronic_002-055-050.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_10", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_55", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21747 + }, + { + "path": "reed_acoustic_018-070-100.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_70", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21748 + }, + { + "path": "organ_electronic_113-056-075.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_24", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_56", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21749 + }, + { + "path": "guitar_acoustic_010-088-100.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_90", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21750 + }, + { + "path": "bass_electronic_027-046-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_85", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21751 + }, + { + "path": "brass_acoustic_046-084-075.wav", + "question": "What numerical value is the MIDI pitch for this note?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_113", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21752 + }, + { + "path": "bass_electronic_027-029-075.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_69", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21753 + }, + { + "path": "organ_electronic_028-028-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_73", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21754 + }, + { + "path": "mallet_acoustic_047-082-050.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_108", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21755 + }, + { + "path": "bass_synthetic_134-052-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_52", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_83", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21756 + }, + { + "path": "bass_synthetic_033-033-025.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_9", + "choice_c": "midi_pitch_21", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_33", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21757 + }, + { + "path": "bass_synthetic_034-046-025.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_70", + "choice_c": "midi_pitch_46", + "choice_d": "midi_pitch_11", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21758 + }, + { + "path": "bass_electronic_018-041-100.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_89", + "choice_b": "midi_pitch_19", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21759 + }, + { + "path": "keyboard_electronic_001-049-127.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_40", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_49", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21760 + }, + { + "path": "bass_synthetic_134-044-025.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21761 + }, + { + "path": "organ_electronic_104-064-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_64", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_64", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21762 + }, + { + "path": "keyboard_electronic_078-041-075.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_79", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21763 + }, + { + "path": "bass_synthetic_068-084-050.wav", + "question": "Select the MIDI pitch that represents this note.", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_40", + "choice_d": "midi_pitch_70", + "answer_gt": "midi_pitch_84", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21764 + }, + { + "path": "organ_electronic_104-034-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_20", + "choice_d": "midi_pitch_34", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21765 + }, + { + "path": "reed_acoustic_011-062-050.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_11", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_96", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21766 + }, + { + "path": "guitar_acoustic_030-062-127.wav", + "question": "The note's sound equates to which MIDI pitch?", + "choice_a": "midi_pitch_62", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_81", + "choice_d": "midi_pitch_88", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21767 + }, + { + "path": "reed_acoustic_037-034-127.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_34", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_105", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21768 + }, + { + "path": "guitar_acoustic_021-077-100.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21769 + }, + { + "path": "string_acoustic_057-082-050.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21770 + }, + { + "path": "bass_synthetic_098-060-100.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_60", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_85", + "choice_d": "midi_pitch_107", + "answer_gt": "midi_pitch_60", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21771 + }, + { + "path": "string_acoustic_056-073-100.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_33", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_73", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21772 + }, + { + "path": "bass_synthetic_009-021-050.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_75", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21773 + }, + { + "path": "guitar_acoustic_021-047-075.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_47", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_22", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21774 + }, + { + "path": "keyboard_acoustic_004-051-050.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_51", + "answer_gt": "midi_pitch_51", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21775 + }, + { + "path": "bass_synthetic_135-082-127.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_50", + "choice_c": "midi_pitch_71", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21776 + }, + { + "path": "keyboard_electronic_002-027-025.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21777 + }, + { + "path": "keyboard_synthetic_000-039-075.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_15", + "choice_b": "midi_pitch_76", + "choice_c": "midi_pitch_69", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21778 + }, + { + "path": "flute_synthetic_000-071-075.wav", + "question": "This note corresponds to what MIDI pitch value?", + "choice_a": "midi_pitch_33", + "choice_b": "midi_pitch_60", + "choice_c": "midi_pitch_82", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21779 + }, + { + "path": "bass_synthetic_009-091-050.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_85", + "choice_b": "midi_pitch_91", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_36", + "answer_gt": "midi_pitch_91", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21780 + }, + { + "path": "keyboard_electronic_001-031-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_10", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21781 + }, + { + "path": "reed_acoustic_023-095-025.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_72", + "answer_gt": "midi_pitch_95", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21782 + }, + { + "path": "bass_electronic_018-039-025.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_26", + "choice_b": "midi_pitch_100", + "choice_c": "midi_pitch_39", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21783 + }, + { + "path": "string_acoustic_014-025-075.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_97", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21784 + }, + { + "path": "organ_electronic_113-081-100.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_107", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_81", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21785 + }, + { + "path": "keyboard_electronic_002-074-075.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_74", + "choice_b": "midi_pitch_79", + "choice_c": "midi_pitch_64", + "choice_d": "midi_pitch_102", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21786 + }, + { + "path": "keyboard_synthetic_000-082-025.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_98", + "choice_b": "midi_pitch_82", + "choice_c": "midi_pitch_31", + "choice_d": "midi_pitch_49", + "answer_gt": "midi_pitch_82", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21787 + }, + { + "path": "bass_synthetic_135-044-050.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_84", + "answer_gt": "midi_pitch_44", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21788 + }, + { + "path": "brass_acoustic_006-087-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_87", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21789 + }, + { + "path": "bass_synthetic_034-045-025.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_116", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21790 + }, + { + "path": "flute_synthetic_000-074-050.wav", + "question": "The single musical note has a MIDI pitch of what?", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_74", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21791 + }, + { + "path": "guitar_acoustic_021-099-050.wav", + "question": "What MIDI pitch does this note translate to?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_99", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21792 + }, + { + "path": "string_acoustic_012-029-127.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_91", + "choice_b": "midi_pitch_87", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21793 + }, + { + "path": "bass_synthetic_134-065-127.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_65", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_73", + "choice_d": "midi_pitch_41", + "answer_gt": "midi_pitch_65", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21794 + }, + { + "path": "organ_electronic_001-075-050.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_104", + "choice_b": "midi_pitch_101", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21795 + }, + { + "path": "guitar_acoustic_010-100-100.wav", + "question": "Identify the MIDI pitch that this note falls under.", + "choice_a": "midi_pitch_36", + "choice_b": "midi_pitch_38", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_92", + "answer_gt": "midi_pitch_100", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21796 + }, + { + "path": "organ_electronic_028-042-025.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_107", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_97", + "choice_d": "midi_pitch_42", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21797 + }, + { + "path": "organ_electronic_007-078-050.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_103", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_78", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21798 + }, + { + "path": "keyboard_electronic_002-075-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_99", + "choice_c": "midi_pitch_75", + "choice_d": "midi_pitch_24", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21799 + }, + { + "path": "mallet_acoustic_047-098-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_32", + "choice_c": "midi_pitch_30", + "choice_d": "midi_pitch_98", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21800 + }, + { + "path": "flute_synthetic_000-057-100.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_77", + "choice_b": "midi_pitch_106", + "choice_c": "midi_pitch_57", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21801 + }, + { + "path": "organ_electronic_001-059-075.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_12", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_108", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21802 + }, + { + "path": "bass_synthetic_033-063-075.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_57", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_63", + "answer_gt": "midi_pitch_63", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21803 + }, + { + "path": "guitar_electronic_022-077-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_37", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_77", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21804 + }, + { + "path": "bass_electronic_018-029-050.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_84", + "choice_b": "midi_pitch_56", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21805 + }, + { + "path": "guitar_acoustic_015-106-127.wav", + "question": "The note at this timestamp has a MIDI pitch of what?", + "choice_a": "midi_pitch_62", + "choice_b": "midi_pitch_20", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_106", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21806 + }, + { + "path": "guitar_acoustic_021-086-127.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_70", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_103", + "answer_gt": "midi_pitch_86", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21807 + }, + { + "path": "organ_electronic_028-071-127.wav", + "question": "What is the correct MIDI pitch of the note you're hearing?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21808 + }, + { + "path": "guitar_acoustic_021-061-127.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21809 + }, + { + "path": "bass_synthetic_009-015-050.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_19", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_34", + "choice_d": "midi_pitch_15", + "answer_gt": "midi_pitch_15", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21810 + }, + { + "path": "reed_acoustic_011-041-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_47", + "choice_c": "midi_pitch_41", + "choice_d": "midi_pitch_95", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21811 + }, + { + "path": "guitar_acoustic_021-072-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_23", + "choice_d": "midi_pitch_60", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21812 + }, + { + "path": "bass_synthetic_034-026-050.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_23", + "choice_c": "midi_pitch_26", + "choice_d": "midi_pitch_16", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21813 + }, + { + "path": "vocal_synthetic_003-061-127.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_21", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_84", + "choice_d": "midi_pitch_82", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21814 + }, + { + "path": "bass_synthetic_098-104-050.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_15", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21815 + }, + { + "path": "organ_electronic_057-062-100.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_35", + "choice_d": "midi_pitch_62", + "answer_gt": "midi_pitch_62", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21816 + }, + { + "path": "bass_synthetic_034-021-050.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_21", + "choice_c": "midi_pitch_88", + "choice_d": "midi_pitch_22", + "answer_gt": "midi_pitch_21", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21817 + }, + { + "path": "organ_electronic_028-093-075.wav", + "question": "The note in this music piece has a MIDI pitch of what?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_24", + "choice_c": "midi_pitch_72", + "choice_d": "midi_pitch_21", + "answer_gt": "midi_pitch_93", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21818 + }, + { + "path": "mallet_acoustic_047-103-025.wav", + "question": "What is the MIDI pitch of this note at the current moment?", + "choice_a": "midi_pitch_63", + "choice_b": "midi_pitch_103", + "choice_c": "midi_pitch_78", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_103", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21819 + }, + { + "path": "reed_acoustic_011-037-050.wav", + "question": "The MIDI pitch for the lone note in this music is what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_37", + "choice_d": "midi_pitch_38", + "answer_gt": "midi_pitch_37", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21820 + }, + { + "path": "organ_electronic_113-043-100.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_105", + "choice_d": "midi_pitch_9", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21821 + }, + { + "path": "keyboard_acoustic_004-045-100.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_75", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_86", + "choice_d": "midi_pitch_45", + "answer_gt": "midi_pitch_45", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21822 + }, + { + "path": "brass_acoustic_046-052-075.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_116", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_52", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21823 + }, + { + "path": "mallet_acoustic_062-042-075.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_58", + "choice_b": "midi_pitch_27", + "choice_c": "midi_pitch_42", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21824 + }, + { + "path": "keyboard_acoustic_004-098-127.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_9", + "choice_b": "midi_pitch_81", + "choice_c": "midi_pitch_98", + "choice_d": "midi_pitch_29", + "answer_gt": "midi_pitch_98", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21825 + }, + { + "path": "vocal_acoustic_000-068-075.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_68", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_11", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_68", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21826 + }, + { + "path": "guitar_acoustic_021-102-100.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_42", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_102", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_102", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21827 + }, + { + "path": "reed_acoustic_023-080-075.wav", + "question": "The specific MIDI pitch for this note is what?", + "choice_a": "midi_pitch_49", + "choice_b": "midi_pitch_45", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21828 + }, + { + "path": "organ_electronic_104-075-100.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_62", + "choice_c": "midi_pitch_10", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_75", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21829 + }, + { + "path": "mallet_acoustic_056-043-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_43", + "choice_b": "midi_pitch_93", + "choice_c": "midi_pitch_66", + "choice_d": "midi_pitch_52", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21830 + }, + { + "path": "guitar_acoustic_010-025-100.wav", + "question": "The MIDI pitch for this note can be identified as what?", + "choice_a": "midi_pitch_83", + "choice_b": "midi_pitch_71", + "choice_c": "midi_pitch_25", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_25", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21831 + }, + { + "path": "bass_synthetic_034-034-025.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_34", + "choice_b": "midi_pitch_51", + "choice_c": "midi_pitch_16", + "choice_d": "midi_pitch_75", + "answer_gt": "midi_pitch_34", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21832 + }, + { + "path": "organ_electronic_104-071-025.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_116", + "choice_c": "midi_pitch_17", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_71", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21833 + }, + { + "path": "organ_electronic_028-054-127.wav", + "question": "The MIDI pitch for this musical tone is what?", + "choice_a": "midi_pitch_59", + "choice_b": "midi_pitch_54", + "choice_c": "midi_pitch_43", + "choice_d": "midi_pitch_48", + "answer_gt": "midi_pitch_54", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21834 + }, + { + "path": "keyboard_synthetic_000-023-075.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_105", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_23", + "answer_gt": "midi_pitch_23", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21835 + }, + { + "path": "mallet_acoustic_047-088-025.wav", + "question": "What is the likely MIDI pitch of the note?", + "choice_a": "midi_pitch_96", + "choice_b": "midi_pitch_88", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21836 + }, + { + "path": "bass_synthetic_068-030-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_44", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_54", + "choice_d": "midi_pitch_106", + "answer_gt": "midi_pitch_30", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21837 + }, + { + "path": "guitar_acoustic_015-043-127.wav", + "question": "This musical note's MIDI pitch is what?", + "choice_a": "midi_pitch_45", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_97", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21838 + }, + { + "path": "string_acoustic_056-041-127.wav", + "question": "The precise MIDI pitch for this note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_41", + "choice_c": "midi_pitch_55", + "choice_d": "midi_pitch_10", + "answer_gt": "midi_pitch_41", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21839 + }, + { + "path": "reed_acoustic_037-042-025.wav", + "question": "This note has a MIDI pitch of what number?", + "choice_a": "midi_pitch_48", + "choice_b": "midi_pitch_42", + "choice_c": "midi_pitch_100", + "choice_d": "midi_pitch_74", + "answer_gt": "midi_pitch_42", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21840 + }, + { + "path": "flute_synthetic_000-026-025.wav", + "question": "The MIDI pitch of this note is identified as what?", + "choice_a": "midi_pitch_46", + "choice_b": "midi_pitch_29", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_26", + "answer_gt": "midi_pitch_26", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21841 + }, + { + "path": "string_acoustic_080-032-075.wav", + "question": "Which MIDI pitch corresponds to the note played?", + "choice_a": "midi_pitch_56", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_68", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21842 + }, + { + "path": "brass_acoustic_016-080-100.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_80", + "choice_b": "midi_pitch_72", + "choice_c": "midi_pitch_116", + "choice_d": "midi_pitch_93", + "answer_gt": "midi_pitch_80", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21843 + }, + { + "path": "guitar_acoustic_021-059-025.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_93", + "choice_b": "midi_pitch_59", + "choice_c": "midi_pitch_108", + "choice_d": "midi_pitch_13", + "answer_gt": "midi_pitch_59", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21844 + }, + { + "path": "keyboard_synthetic_000-046-025.wav", + "question": "The MIDI pitch of the note being heard is what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_77", + "choice_c": "midi_pitch_60", + "choice_d": "midi_pitch_46", + "answer_gt": "midi_pitch_46", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21845 + }, + { + "path": "bass_synthetic_033-027-050.wav", + "question": "The MIDI pitch of this distinct note is what?", + "choice_a": "midi_pitch_27", + "choice_b": "midi_pitch_30", + "choice_c": "midi_pitch_61", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_27", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21846 + }, + { + "path": "keyboard_electronic_069-077-025.wav", + "question": "The sound of this note corresponds to which MIDI pitch?", + "choice_a": "midi_pitch_113", + "choice_b": "midi_pitch_17", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_77", + "answer_gt": "midi_pitch_77", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21847 + }, + { + "path": "bass_synthetic_009-035-025.wav", + "question": "This isolated note's MIDI pitch translates to what?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_35", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_100", + "answer_gt": "midi_pitch_35", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21848 + }, + { + "path": "string_acoustic_057-066-100.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_66", + "choice_b": "midi_pitch_80", + "choice_c": "midi_pitch_44", + "choice_d": "midi_pitch_81", + "answer_gt": "midi_pitch_66", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21849 + }, + { + "path": "guitar_acoustic_010-031-100.wav", + "question": "The MIDI pitch of the musical note is identified as what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_86", + "choice_c": "midi_pitch_33", + "choice_d": "midi_pitch_43", + "answer_gt": "midi_pitch_31", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21850 + }, + { + "path": "bass_electronic_018-047-100.wav", + "question": "The MIDI pitch value for this music note is what?", + "choice_a": "midi_pitch_31", + "choice_b": "midi_pitch_55", + "choice_c": "midi_pitch_47", + "choice_d": "midi_pitch_19", + "answer_gt": "midi_pitch_47", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21851 + }, + { + "path": "brass_acoustic_006-032-025.wav", + "question": "The MIDI pitch that denotes this note is what?", + "choice_a": "midi_pitch_79", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_87", + "choice_d": "midi_pitch_32", + "answer_gt": "midi_pitch_32", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21852 + }, + { + "path": "guitar_acoustic_014-061-100.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_61", + "choice_b": "midi_pitch_31", + "choice_c": "midi_pitch_94", + "choice_d": "midi_pitch_71", + "answer_gt": "midi_pitch_61", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21853 + }, + { + "path": "bass_electronic_018-050-127.wav", + "question": "The MIDI pitch number for the note heard is what?", + "choice_a": "midi_pitch_50", + "choice_b": "midi_pitch_36", + "choice_c": "midi_pitch_95", + "choice_d": "midi_pitch_33", + "answer_gt": "midi_pitch_50", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21854 + }, + { + "path": "brass_acoustic_059-039-075.wav", + "question": "Determine the MIDI pitch for the note in question.", + "choice_a": "midi_pitch_94", + "choice_b": "midi_pitch_44", + "choice_c": "midi_pitch_99", + "choice_d": "midi_pitch_39", + "answer_gt": "midi_pitch_39", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21855 + }, + { + "path": "organ_electronic_028-029-025.wav", + "question": "What is the MIDI pitch number for this note's sound?", + "choice_a": "midi_pitch_29", + "choice_b": "midi_pitch_102", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_20", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21856 + }, + { + "path": "guitar_acoustic_010-022-050.wav", + "question": "What MIDI pitch does this note's frequency correspond to?", + "choice_a": "midi_pitch_71", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_59", + "choice_d": "midi_pitch_99", + "answer_gt": "midi_pitch_22", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21857 + }, + { + "path": "organ_electronic_001-078-025.wav", + "question": "This single note registers as which MIDI pitch?", + "choice_a": "midi_pitch_78", + "choice_b": "midi_pitch_65", + "choice_c": "midi_pitch_67", + "choice_d": "midi_pitch_14", + "answer_gt": "midi_pitch_78", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21858 + }, + { + "path": "bass_electronic_027-058-075.wav", + "question": "This note's frequency translates to which MIDI pitch?", + "choice_a": "midi_pitch_28", + "choice_b": "midi_pitch_58", + "choice_c": "midi_pitch_103", + "choice_d": "midi_pitch_87", + "answer_gt": "midi_pitch_58", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21859 + }, + { + "path": "string_acoustic_080-028-025.wav", + "question": "What MIDI pitch is the note equivalent to?", + "choice_a": "midi_pitch_90", + "choice_b": "midi_pitch_28", + "choice_c": "midi_pitch_12", + "choice_d": "midi_pitch_37", + "answer_gt": "midi_pitch_28", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21860 + }, + { + "path": "bass_electronic_025-036-050.wav", + "question": "What is the MIDI pitch value of this isolated note?", + "choice_a": "midi_pitch_25", + "choice_b": "midi_pitch_78", + "choice_c": "midi_pitch_36", + "choice_d": "midi_pitch_76", + "answer_gt": "midi_pitch_36", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21861 + }, + { + "path": "keyboard_electronic_002-094-025.wav", + "question": "The MIDI pitch associated with this note is what?", + "choice_a": "midi_pitch_32", + "choice_b": "midi_pitch_94", + "choice_c": "midi_pitch_113", + "choice_d": "midi_pitch_101", + "answer_gt": "midi_pitch_94", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21862 + }, + { + "path": "guitar_electronic_022-043-050.wav", + "question": "Identify the MIDI pitch that matches this note's frequency.", + "choice_a": "midi_pitch_92", + "choice_b": "midi_pitch_43", + "choice_c": "midi_pitch_101", + "choice_d": "midi_pitch_40", + "answer_gt": "midi_pitch_43", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21863 + }, + { + "path": "reed_acoustic_023-076-100.wav", + "question": "What is the MIDI pitch assigned to this musical sound?", + "choice_a": "midi_pitch_76", + "choice_b": "midi_pitch_14", + "choice_c": "midi_pitch_93", + "choice_d": "midi_pitch_56", + "answer_gt": "midi_pitch_76", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21864 + }, + { + "path": "guitar_acoustic_010-053-075.wav", + "question": "The MIDI pitch for this note is determined as what?", + "choice_a": "midi_pitch_23", + "choice_b": "midi_pitch_22", + "choice_c": "midi_pitch_53", + "choice_d": "midi_pitch_28", + "answer_gt": "midi_pitch_53", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21865 + }, + { + "path": "keyboard_electronic_001-029-100.wav", + "question": "What MIDI pitch is being represented by this note?", + "choice_a": "midi_pitch_16", + "choice_b": "midi_pitch_107", + "choice_c": "midi_pitch_29", + "choice_d": "midi_pitch_12", + "answer_gt": "midi_pitch_29", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21866 + }, + { + "path": "keyboard_acoustic_004-085-075.wav", + "question": "The MIDI pitch designation for this note is what?", + "choice_a": "midi_pitch_53", + "choice_b": "midi_pitch_85", + "choice_c": "midi_pitch_83", + "choice_d": "midi_pitch_94", + "answer_gt": "midi_pitch_85", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21867 + }, + { + "path": "guitar_electronic_022-088-127.wav", + "question": "The MIDI pitch corresponding to this note is what?", + "choice_a": "midi_pitch_88", + "choice_b": "midi_pitch_61", + "choice_c": "midi_pitch_80", + "choice_d": "midi_pitch_89", + "answer_gt": "midi_pitch_88", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21868 + }, + { + "path": "bass_synthetic_134-057-075.wav", + "question": "The MIDI pitch of this music note is what value?", + "choice_a": "midi_pitch_12", + "choice_b": "midi_pitch_108", + "choice_c": "midi_pitch_13", + "choice_d": "midi_pitch_57", + "answer_gt": "midi_pitch_57", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21869 + }, + { + "path": "mallet_acoustic_062-072-075.wav", + "question": "What is the MIDI pitch level of the note played?", + "choice_a": "midi_pitch_72", + "choice_b": "midi_pitch_92", + "choice_c": "midi_pitch_52", + "choice_d": "midi_pitch_67", + "answer_gt": "midi_pitch_72", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21870 + }, + { + "path": "bass_synthetic_068-104-127.wav", + "question": "The MIDI pitch for this note falls at what number?", + "choice_a": "midi_pitch_51", + "choice_b": "midi_pitch_104", + "choice_c": "midi_pitch_62", + "choice_d": "midi_pitch_113", + "answer_gt": "midi_pitch_104", + "task_name": "Music_Midi_Pitch_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21871 + }, + { + "path": "flute_acoustic_002-103-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21872 + }, + { + "path": "bass_electronic_025-059-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21873 + }, + { + "path": "keyboard_electronic_002-033-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21874 + }, + { + "path": "guitar_acoustic_014-067-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21875 + }, + { + "path": "keyboard_synthetic_000-063-127.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21876 + }, + { + "path": "guitar_acoustic_030-064-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21877 + }, + { + "path": "bass_synthetic_009-044-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21878 + }, + { + "path": "bass_synthetic_033-105-050.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21879 + }, + { + "path": "keyboard_electronic_078-061-025.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21880 + }, + { + "path": "guitar_acoustic_010-027-100.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21881 + }, + { + "path": "mallet_acoustic_062-107-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21882 + }, + { + "path": "keyboard_electronic_001-037-075.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21883 + }, + { + "path": "brass_acoustic_015-045-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21884 + }, + { + "path": "mallet_acoustic_056-088-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21885 + }, + { + "path": "bass_electronic_018-043-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21886 + }, + { + "path": "flute_synthetic_000-101-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21887 + }, + { + "path": "organ_electronic_007-096-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21888 + }, + { + "path": "organ_electronic_057-050-025.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21889 + }, + { + "path": "keyboard_electronic_003-047-025.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21890 + }, + { + "path": "reed_acoustic_018-072-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21891 + }, + { + "path": "bass_synthetic_068-084-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21892 + }, + { + "path": "keyboard_electronic_069-068-075.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21893 + }, + { + "path": "bass_synthetic_033-065-100.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21894 + }, + { + "path": "keyboard_electronic_001-024-100.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21895 + }, + { + "path": "organ_electronic_028-068-100.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21896 + }, + { + "path": "keyboard_electronic_078-082-100.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21897 + }, + { + "path": "vocal_synthetic_003-081-100.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21898 + }, + { + "path": "organ_electronic_057-025-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21899 + }, + { + "path": "keyboard_electronic_001-107-025.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21900 + }, + { + "path": "organ_electronic_007-038-075.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21901 + }, + { + "path": "bass_synthetic_134-096-100.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21902 + }, + { + "path": "reed_acoustic_023-097-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21903 + }, + { + "path": "guitar_electronic_028-056-025.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21904 + }, + { + "path": "organ_electronic_007-012-050.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21905 + }, + { + "path": "flute_synthetic_000-037-075.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21906 + }, + { + "path": "reed_acoustic_023-054-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21907 + }, + { + "path": "bass_synthetic_098-021-100.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21908 + }, + { + "path": "flute_synthetic_000-096-050.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21909 + }, + { + "path": "organ_electronic_001-058-075.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21910 + }, + { + "path": "guitar_acoustic_021-055-100.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21911 + }, + { + "path": "keyboard_electronic_069-052-025.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21912 + }, + { + "path": "mallet_acoustic_047-085-050.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21913 + }, + { + "path": "vocal_synthetic_003-096-100.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21914 + }, + { + "path": "keyboard_acoustic_004-079-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21915 + }, + { + "path": "brass_acoustic_006-032-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21916 + }, + { + "path": "keyboard_acoustic_004-049-050.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21917 + }, + { + "path": "keyboard_acoustic_004-025-100.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21918 + }, + { + "path": "bass_synthetic_009-033-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21919 + }, + { + "path": "guitar_electronic_022-053-075.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21920 + }, + { + "path": "organ_electronic_028-106-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21921 + }, + { + "path": "organ_electronic_057-068-127.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21922 + }, + { + "path": "reed_acoustic_037-073-100.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21923 + }, + { + "path": "string_acoustic_080-034-127.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21924 + }, + { + "path": "brass_acoustic_059-050-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21925 + }, + { + "path": "reed_acoustic_037-057-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21926 + }, + { + "path": "reed_acoustic_037-055-075.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21927 + }, + { + "path": "guitar_electronic_022-040-100.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21928 + }, + { + "path": "flute_acoustic_002-069-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21929 + }, + { + "path": "organ_electronic_028-081-050.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21930 + }, + { + "path": "organ_electronic_113-065-050.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21931 + }, + { + "path": "bass_electronic_025-030-025.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21932 + }, + { + "path": "keyboard_electronic_002-084-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21933 + }, + { + "path": "flute_acoustic_002-107-127.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21934 + }, + { + "path": "organ_electronic_028-027-127.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21935 + }, + { + "path": "guitar_acoustic_021-035-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21936 + }, + { + "path": "keyboard_synthetic_000-050-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21937 + }, + { + "path": "bass_synthetic_068-075-127.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21938 + }, + { + "path": "guitar_acoustic_021-040-075.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21939 + }, + { + "path": "organ_electronic_007-022-100.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21940 + }, + { + "path": "bass_synthetic_033-050-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21941 + }, + { + "path": "bass_synthetic_098-083-100.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21942 + }, + { + "path": "vocal_synthetic_003-068-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21943 + }, + { + "path": "bass_synthetic_134-039-127.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21944 + }, + { + "path": "keyboard_acoustic_004-103-100.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21945 + }, + { + "path": "keyboard_synthetic_000-073-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21946 + }, + { + "path": "brass_acoustic_016-084-075.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21947 + }, + { + "path": "string_acoustic_057-065-100.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21948 + }, + { + "path": "bass_synthetic_009-091-100.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21949 + }, + { + "path": "bass_synthetic_134-044-100.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21950 + }, + { + "path": "reed_acoustic_023-059-100.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21951 + }, + { + "path": "keyboard_electronic_098-034-075.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21952 + }, + { + "path": "organ_electronic_113-047-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21953 + }, + { + "path": "organ_electronic_057-053-100.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21954 + }, + { + "path": "guitar_acoustic_015-071-050.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21955 + }, + { + "path": "keyboard_acoustic_004-069-127.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21956 + }, + { + "path": "keyboard_electronic_003-058-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21957 + }, + { + "path": "keyboard_synthetic_000-051-100.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21958 + }, + { + "path": "string_acoustic_056-064-025.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21959 + }, + { + "path": "bass_synthetic_009-009-050.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21960 + }, + { + "path": "guitar_acoustic_014-043-075.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21961 + }, + { + "path": "guitar_electronic_028-050-100.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21962 + }, + { + "path": "string_acoustic_071-038-127.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21963 + }, + { + "path": "guitar_acoustic_010-043-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21964 + }, + { + "path": "vocal_synthetic_003-103-100.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21965 + }, + { + "path": "keyboard_electronic_078-069-025.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21966 + }, + { + "path": "guitar_acoustic_010-071-025.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21967 + }, + { + "path": "keyboard_electronic_002-033-075.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21968 + }, + { + "path": "bass_electronic_027-031-025.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21969 + }, + { + "path": "keyboard_electronic_003-088-075.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21970 + }, + { + "path": "keyboard_electronic_002-047-075.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21971 + }, + { + "path": "bass_synthetic_068-036-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21972 + }, + { + "path": "reed_acoustic_018-086-127.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21973 + }, + { + "path": "flute_acoustic_002-081-025.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21974 + }, + { + "path": "guitar_acoustic_014-106-100.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21975 + }, + { + "path": "keyboard_acoustic_004-056-075.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21976 + }, + { + "path": "guitar_acoustic_030-058-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21977 + }, + { + "path": "guitar_acoustic_015-075-100.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21978 + }, + { + "path": "reed_acoustic_011-057-050.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21979 + }, + { + "path": "bass_electronic_027-035-100.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21980 + }, + { + "path": "bass_synthetic_034-102-025.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21981 + }, + { + "path": "guitar_acoustic_021-082-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21982 + }, + { + "path": "organ_electronic_007-096-025.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21983 + }, + { + "path": "organ_electronic_113-045-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21984 + }, + { + "path": "guitar_acoustic_015-102-050.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21985 + }, + { + "path": "vocal_synthetic_003-093-100.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21986 + }, + { + "path": "keyboard_electronic_001-031-100.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21987 + }, + { + "path": "keyboard_electronic_003-028-050.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21988 + }, + { + "path": "bass_electronic_027-023-025.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21989 + }, + { + "path": "bass_synthetic_068-057-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21990 + }, + { + "path": "keyboard_electronic_098-023-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21991 + }, + { + "path": "bass_synthetic_068-038-025.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21992 + }, + { + "path": "keyboard_electronic_002-095-127.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21993 + }, + { + "path": "mallet_acoustic_047-089-050.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21994 + }, + { + "path": "keyboard_electronic_069-046-025.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21995 + }, + { + "path": "organ_electronic_057-072-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21996 + }, + { + "path": "bass_synthetic_033-044-050.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21997 + }, + { + "path": "organ_electronic_007-064-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21998 + }, + { + "path": "bass_synthetic_068-049-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 21999 + }, + { + "path": "keyboard_electronic_098-065-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22000 + }, + { + "path": "brass_acoustic_016-088-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22001 + }, + { + "path": "guitar_electronic_022-072-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22002 + }, + { + "path": "keyboard_electronic_001-034-100.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22003 + }, + { + "path": "bass_synthetic_009-034-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22004 + }, + { + "path": "brass_acoustic_046-071-050.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22005 + }, + { + "path": "bass_synthetic_135-035-025.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22006 + }, + { + "path": "bass_synthetic_098-045-075.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22007 + }, + { + "path": "guitar_acoustic_014-067-025.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22008 + }, + { + "path": "guitar_acoustic_014-073-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22009 + }, + { + "path": "keyboard_electronic_001-103-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22010 + }, + { + "path": "keyboard_electronic_098-058-025.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22011 + }, + { + "path": "mallet_acoustic_047-091-100.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22012 + }, + { + "path": "mallet_acoustic_056-116-127.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22013 + }, + { + "path": "keyboard_electronic_069-032-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22014 + }, + { + "path": "reed_acoustic_037-062-050.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22015 + }, + { + "path": "organ_electronic_057-026-075.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22016 + }, + { + "path": "keyboard_acoustic_004-038-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22017 + }, + { + "path": "guitar_acoustic_030-045-050.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22018 + }, + { + "path": "guitar_acoustic_021-069-127.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22019 + }, + { + "path": "flute_synthetic_000-070-075.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22020 + }, + { + "path": "brass_acoustic_016-078-025.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22021 + }, + { + "path": "guitar_acoustic_010-058-075.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22022 + }, + { + "path": "guitar_acoustic_030-097-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22023 + }, + { + "path": "keyboard_acoustic_004-104-100.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22024 + }, + { + "path": "bass_synthetic_033-073-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22025 + }, + { + "path": "guitar_acoustic_015-067-075.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22026 + }, + { + "path": "keyboard_synthetic_000-081-100.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22027 + }, + { + "path": "keyboard_electronic_078-060-025.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22028 + }, + { + "path": "mallet_acoustic_062-061-050.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22029 + }, + { + "path": "flute_synthetic_000-095-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22030 + }, + { + "path": "brass_acoustic_059-040-025.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22031 + }, + { + "path": "mallet_acoustic_047-104-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22032 + }, + { + "path": "bass_synthetic_068-067-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22033 + }, + { + "path": "guitar_acoustic_010-054-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22034 + }, + { + "path": "organ_electronic_057-079-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22035 + }, + { + "path": "brass_acoustic_006-073-075.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22036 + }, + { + "path": "bass_synthetic_135-096-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22037 + }, + { + "path": "keyboard_synthetic_000-079-075.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22038 + }, + { + "path": "bass_synthetic_135-079-075.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22039 + }, + { + "path": "bass_electronic_027-032-127.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22040 + }, + { + "path": "keyboard_synthetic_000-021-075.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22041 + }, + { + "path": "guitar_electronic_028-087-075.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22042 + }, + { + "path": "organ_electronic_007-082-127.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22043 + }, + { + "path": "mallet_acoustic_062-062-025.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22044 + }, + { + "path": "keyboard_acoustic_004-050-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22045 + }, + { + "path": "keyboard_electronic_098-024-075.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22046 + }, + { + "path": "organ_electronic_007-048-025.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22047 + }, + { + "path": "vocal_synthetic_003-048-100.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22048 + }, + { + "path": "bass_synthetic_034-060-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22049 + }, + { + "path": "bass_synthetic_009-092-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22050 + }, + { + "path": "reed_acoustic_023-046-050.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22051 + }, + { + "path": "keyboard_electronic_098-099-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22052 + }, + { + "path": "bass_synthetic_034-064-100.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22053 + }, + { + "path": "mallet_acoustic_062-071-025.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22054 + }, + { + "path": "reed_acoustic_023-085-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22055 + }, + { + "path": "reed_acoustic_018-084-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22056 + }, + { + "path": "keyboard_electronic_069-053-100.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22057 + }, + { + "path": "brass_acoustic_016-088-075.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22058 + }, + { + "path": "bass_synthetic_098-065-050.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22059 + }, + { + "path": "guitar_electronic_028-094-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22060 + }, + { + "path": "brass_acoustic_046-081-100.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22061 + }, + { + "path": "bass_synthetic_135-096-025.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22062 + }, + { + "path": "keyboard_acoustic_004-089-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22063 + }, + { + "path": "string_acoustic_071-044-050.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22064 + }, + { + "path": "bass_synthetic_033-042-025.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22065 + }, + { + "path": "bass_synthetic_134-026-127.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22066 + }, + { + "path": "brass_acoustic_059-063-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22067 + }, + { + "path": "keyboard_synthetic_000-040-075.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22068 + }, + { + "path": "organ_electronic_007-077-127.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22069 + }, + { + "path": "bass_synthetic_135-070-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22070 + }, + { + "path": "bass_electronic_027-038-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22071 + }, + { + "path": "bass_electronic_025-036-127.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22072 + }, + { + "path": "guitar_acoustic_014-088-050.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22073 + }, + { + "path": "keyboard_acoustic_004-057-127.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22074 + }, + { + "path": "reed_acoustic_018-082-050.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22075 + }, + { + "path": "guitar_electronic_028-106-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22076 + }, + { + "path": "bass_synthetic_135-039-100.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22077 + }, + { + "path": "vocal_synthetic_003-067-100.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22078 + }, + { + "path": "brass_acoustic_006-054-127.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22079 + }, + { + "path": "keyboard_electronic_003-046-127.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22080 + }, + { + "path": "keyboard_electronic_001-064-050.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22081 + }, + { + "path": "bass_synthetic_033-077-075.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22082 + }, + { + "path": "mallet_acoustic_062-052-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22083 + }, + { + "path": "brass_acoustic_046-062-025.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22084 + }, + { + "path": "mallet_acoustic_062-038-025.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22085 + }, + { + "path": "keyboard_electronic_098-071-127.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22086 + }, + { + "path": "bass_electronic_018-060-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22087 + }, + { + "path": "keyboard_electronic_069-060-127.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22088 + }, + { + "path": "bass_synthetic_009-045-025.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22089 + }, + { + "path": "bass_electronic_025-060-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22090 + }, + { + "path": "keyboard_electronic_001-072-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22091 + }, + { + "path": "bass_synthetic_134-069-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22092 + }, + { + "path": "bass_synthetic_009-014-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22093 + }, + { + "path": "bass_synthetic_034-023-100.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22094 + }, + { + "path": "brass_acoustic_015-070-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22095 + }, + { + "path": "organ_electronic_104-068-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22096 + }, + { + "path": "bass_synthetic_068-037-025.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22097 + }, + { + "path": "bass_synthetic_098-050-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22098 + }, + { + "path": "vocal_synthetic_003-021-127.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22099 + }, + { + "path": "reed_acoustic_023-097-127.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22100 + }, + { + "path": "bass_synthetic_068-045-025.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22101 + }, + { + "path": "bass_synthetic_134-040-025.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22102 + }, + { + "path": "bass_synthetic_134-093-127.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22103 + }, + { + "path": "brass_acoustic_006-065-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22104 + }, + { + "path": "bass_synthetic_135-045-075.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22105 + }, + { + "path": "keyboard_electronic_098-086-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22106 + }, + { + "path": "mallet_acoustic_056-082-100.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22107 + }, + { + "path": "bass_synthetic_098-077-025.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22108 + }, + { + "path": "flute_synthetic_000-088-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22109 + }, + { + "path": "guitar_acoustic_010-051-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22110 + }, + { + "path": "brass_acoustic_046-071-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22111 + }, + { + "path": "organ_electronic_001-040-100.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22112 + }, + { + "path": "bass_synthetic_134-105-050.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22113 + }, + { + "path": "keyboard_electronic_001-022-050.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22114 + }, + { + "path": "reed_acoustic_011-053-100.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22115 + }, + { + "path": "bass_electronic_027-029-050.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22116 + }, + { + "path": "organ_electronic_007-042-025.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22117 + }, + { + "path": "keyboard_electronic_069-048-127.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22118 + }, + { + "path": "bass_synthetic_033-061-127.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22119 + }, + { + "path": "flute_synthetic_000-103-025.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22120 + }, + { + "path": "bass_synthetic_134-050-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22121 + }, + { + "path": "organ_electronic_007-077-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22122 + }, + { + "path": "bass_synthetic_009-044-025.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22123 + }, + { + "path": "guitar_acoustic_010-108-075.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22124 + }, + { + "path": "organ_electronic_028-053-050.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22125 + }, + { + "path": "guitar_electronic_022-059-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22126 + }, + { + "path": "guitar_electronic_028-077-075.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22127 + }, + { + "path": "string_acoustic_080-039-050.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22128 + }, + { + "path": "brass_acoustic_006-068-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22129 + }, + { + "path": "keyboard_electronic_098-038-127.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22130 + }, + { + "path": "guitar_electronic_028-044-025.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22131 + }, + { + "path": "keyboard_electronic_002-084-025.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22132 + }, + { + "path": "mallet_acoustic_062-049-075.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22133 + }, + { + "path": "mallet_acoustic_062-080-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22134 + }, + { + "path": "organ_electronic_113-075-050.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22135 + }, + { + "path": "mallet_acoustic_062-028-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22136 + }, + { + "path": "organ_electronic_104-032-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22137 + }, + { + "path": "mallet_acoustic_056-035-050.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22138 + }, + { + "path": "reed_acoustic_018-063-075.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22139 + }, + { + "path": "string_acoustic_014-041-100.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22140 + }, + { + "path": "string_acoustic_057-061-100.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22141 + }, + { + "path": "reed_acoustic_037-066-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22142 + }, + { + "path": "keyboard_acoustic_004-067-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22143 + }, + { + "path": "mallet_acoustic_062-058-050.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22144 + }, + { + "path": "brass_acoustic_016-068-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22145 + }, + { + "path": "keyboard_acoustic_004-023-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22146 + }, + { + "path": "organ_electronic_104-056-025.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22147 + }, + { + "path": "mallet_acoustic_056-058-050.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22148 + }, + { + "path": "organ_electronic_104-052-025.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22149 + }, + { + "path": "reed_acoustic_018-062-025.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22150 + }, + { + "path": "keyboard_electronic_098-056-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22151 + }, + { + "path": "mallet_acoustic_062-057-050.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22152 + }, + { + "path": "bass_synthetic_098-041-075.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22153 + }, + { + "path": "brass_acoustic_016-056-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22154 + }, + { + "path": "brass_acoustic_046-106-025.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22155 + }, + { + "path": "organ_electronic_007-081-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22156 + }, + { + "path": "flute_synthetic_000-040-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22157 + }, + { + "path": "bass_synthetic_068-106-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22158 + }, + { + "path": "keyboard_electronic_001-040-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22159 + }, + { + "path": "bass_synthetic_068-049-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22160 + }, + { + "path": "string_acoustic_056-050-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22161 + }, + { + "path": "bass_electronic_025-028-050.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22162 + }, + { + "path": "organ_electronic_028-098-025.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22163 + }, + { + "path": "string_acoustic_056-068-025.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22164 + }, + { + "path": "guitar_acoustic_021-076-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22165 + }, + { + "path": "flute_synthetic_000-072-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22166 + }, + { + "path": "keyboard_electronic_001-044-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22167 + }, + { + "path": "keyboard_electronic_098-065-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22168 + }, + { + "path": "bass_synthetic_033-057-100.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22169 + }, + { + "path": "guitar_acoustic_030-085-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22170 + }, + { + "path": "guitar_acoustic_021-091-127.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22171 + }, + { + "path": "bass_synthetic_009-012-075.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22172 + }, + { + "path": "organ_electronic_057-076-127.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22173 + }, + { + "path": "guitar_electronic_028-045-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22174 + }, + { + "path": "keyboard_electronic_098-079-127.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22175 + }, + { + "path": "keyboard_electronic_003-037-050.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22176 + }, + { + "path": "reed_acoustic_018-067-050.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22177 + }, + { + "path": "guitar_electronic_028-096-050.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22178 + }, + { + "path": "organ_electronic_007-051-127.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22179 + }, + { + "path": "keyboard_electronic_001-089-050.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22180 + }, + { + "path": "organ_electronic_001-064-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22181 + }, + { + "path": "guitar_acoustic_014-031-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22182 + }, + { + "path": "reed_acoustic_011-063-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22183 + }, + { + "path": "flute_synthetic_000-074-127.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22184 + }, + { + "path": "mallet_acoustic_062-045-100.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22185 + }, + { + "path": "bass_synthetic_134-108-050.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22186 + }, + { + "path": "keyboard_acoustic_004-040-127.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22187 + }, + { + "path": "reed_acoustic_023-073-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22188 + }, + { + "path": "bass_synthetic_068-032-050.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22189 + }, + { + "path": "keyboard_acoustic_004-046-127.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22190 + }, + { + "path": "keyboard_synthetic_000-032-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22191 + }, + { + "path": "flute_acoustic_002-073-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22192 + }, + { + "path": "bass_synthetic_098-049-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22193 + }, + { + "path": "keyboard_electronic_003-034-050.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22194 + }, + { + "path": "flute_synthetic_000-090-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22195 + }, + { + "path": "mallet_acoustic_062-058-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22196 + }, + { + "path": "reed_acoustic_037-073-025.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22197 + }, + { + "path": "string_acoustic_056-056-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22198 + }, + { + "path": "bass_synthetic_034-094-127.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22199 + }, + { + "path": "mallet_acoustic_047-104-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22200 + }, + { + "path": "string_acoustic_080-028-075.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22201 + }, + { + "path": "keyboard_electronic_098-107-100.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22202 + }, + { + "path": "brass_acoustic_015-077-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22203 + }, + { + "path": "bass_electronic_027-036-100.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22204 + }, + { + "path": "bass_synthetic_135-091-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22205 + }, + { + "path": "organ_electronic_028-088-075.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22206 + }, + { + "path": "organ_electronic_104-038-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22207 + }, + { + "path": "keyboard_electronic_003-072-025.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22208 + }, + { + "path": "vocal_synthetic_003-039-100.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22209 + }, + { + "path": "bass_synthetic_068-070-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22210 + }, + { + "path": "flute_synthetic_000-068-100.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22211 + }, + { + "path": "bass_synthetic_033-099-100.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22212 + }, + { + "path": "keyboard_acoustic_004-037-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22213 + }, + { + "path": "bass_synthetic_135-034-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22214 + }, + { + "path": "keyboard_acoustic_004-077-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22215 + }, + { + "path": "mallet_acoustic_047-092-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22216 + }, + { + "path": "bass_synthetic_033-060-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22217 + }, + { + "path": "guitar_acoustic_010-075-127.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22218 + }, + { + "path": "organ_electronic_057-060-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22219 + }, + { + "path": "bass_synthetic_134-055-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22220 + }, + { + "path": "organ_electronic_104-032-100.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22221 + }, + { + "path": "keyboard_acoustic_004-030-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22222 + }, + { + "path": "guitar_acoustic_014-106-050.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22223 + }, + { + "path": "bass_synthetic_135-073-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22224 + }, + { + "path": "mallet_acoustic_062-095-127.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22225 + }, + { + "path": "bass_synthetic_033-107-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22226 + }, + { + "path": "keyboard_electronic_069-045-100.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22227 + }, + { + "path": "bass_synthetic_009-096-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22228 + }, + { + "path": "organ_electronic_104-041-050.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22229 + }, + { + "path": "brass_acoustic_059-035-100.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22230 + }, + { + "path": "bass_synthetic_033-031-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22231 + }, + { + "path": "flute_synthetic_000-043-075.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22232 + }, + { + "path": "bass_synthetic_033-104-075.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22233 + }, + { + "path": "brass_acoustic_059-032-127.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22234 + }, + { + "path": "flute_synthetic_000-071-127.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22235 + }, + { + "path": "guitar_electronic_028-044-127.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22236 + }, + { + "path": "string_acoustic_056-039-075.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22237 + }, + { + "path": "reed_acoustic_037-075-050.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22238 + }, + { + "path": "bass_synthetic_033-090-025.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22239 + }, + { + "path": "keyboard_electronic_098-079-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22240 + }, + { + "path": "guitar_acoustic_030-041-100.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22241 + }, + { + "path": "reed_acoustic_023-043-075.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22242 + }, + { + "path": "organ_electronic_057-036-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22243 + }, + { + "path": "bass_synthetic_033-025-075.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22244 + }, + { + "path": "organ_electronic_007-044-100.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22245 + }, + { + "path": "string_acoustic_014-062-127.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22246 + }, + { + "path": "flute_acoustic_002-077-050.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22247 + }, + { + "path": "string_acoustic_014-029-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22248 + }, + { + "path": "reed_acoustic_023-071-127.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22249 + }, + { + "path": "keyboard_electronic_003-079-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22250 + }, + { + "path": "keyboard_electronic_001-065-127.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22251 + }, + { + "path": "bass_synthetic_033-089-050.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22252 + }, + { + "path": "reed_acoustic_023-039-050.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22253 + }, + { + "path": "keyboard_electronic_069-078-025.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22254 + }, + { + "path": "keyboard_electronic_098-028-075.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22255 + }, + { + "path": "keyboard_electronic_003-086-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22256 + }, + { + "path": "keyboard_acoustic_004-061-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22257 + }, + { + "path": "bass_synthetic_134-075-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22258 + }, + { + "path": "guitar_acoustic_014-022-025.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22259 + }, + { + "path": "organ_electronic_007-037-025.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22260 + }, + { + "path": "vocal_synthetic_003-091-025.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22261 + }, + { + "path": "reed_acoustic_023-056-050.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22262 + }, + { + "path": "guitar_acoustic_010-059-075.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22263 + }, + { + "path": "flute_synthetic_000-061-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22264 + }, + { + "path": "organ_electronic_104-042-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22265 + }, + { + "path": "flute_acoustic_002-074-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22266 + }, + { + "path": "guitar_electronic_028-101-127.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22267 + }, + { + "path": "bass_synthetic_033-049-127.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22268 + }, + { + "path": "bass_synthetic_135-050-025.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22269 + }, + { + "path": "brass_acoustic_016-080-025.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22270 + }, + { + "path": "string_acoustic_056-075-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22271 + }, + { + "path": "guitar_electronic_028-057-075.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22272 + }, + { + "path": "string_acoustic_056-049-025.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22273 + }, + { + "path": "bass_synthetic_068-027-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22274 + }, + { + "path": "string_acoustic_057-063-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22275 + }, + { + "path": "guitar_acoustic_010-107-075.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22276 + }, + { + "path": "mallet_acoustic_056-083-127.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22277 + }, + { + "path": "organ_electronic_028-040-050.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22278 + }, + { + "path": "guitar_acoustic_014-107-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22279 + }, + { + "path": "guitar_electronic_022-075-127.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22280 + }, + { + "path": "bass_synthetic_034-021-025.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22281 + }, + { + "path": "bass_electronic_027-046-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22282 + }, + { + "path": "flute_synthetic_000-030-075.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22283 + }, + { + "path": "bass_synthetic_098-073-050.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22284 + }, + { + "path": "guitar_acoustic_030-023-025.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22285 + }, + { + "path": "bass_synthetic_033-061-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22286 + }, + { + "path": "flute_synthetic_000-104-050.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22287 + }, + { + "path": "organ_electronic_028-050-100.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22288 + }, + { + "path": "keyboard_electronic_002-103-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22289 + }, + { + "path": "guitar_acoustic_021-056-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22290 + }, + { + "path": "string_acoustic_014-048-050.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22291 + }, + { + "path": "guitar_acoustic_015-075-025.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22292 + }, + { + "path": "organ_electronic_028-106-100.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22293 + }, + { + "path": "bass_synthetic_134-032-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22294 + }, + { + "path": "brass_acoustic_016-057-127.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22295 + }, + { + "path": "keyboard_electronic_001-073-050.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22296 + }, + { + "path": "reed_acoustic_023-068-025.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22297 + }, + { + "path": "bass_synthetic_009-060-075.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22298 + }, + { + "path": "keyboard_electronic_098-066-025.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22299 + }, + { + "path": "organ_electronic_001-095-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22300 + }, + { + "path": "bass_electronic_018-043-050.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22301 + }, + { + "path": "string_acoustic_080-052-100.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22302 + }, + { + "path": "bass_synthetic_068-038-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22303 + }, + { + "path": "guitar_acoustic_015-057-025.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22304 + }, + { + "path": "string_acoustic_014-051-025.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22305 + }, + { + "path": "keyboard_electronic_098-063-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22306 + }, + { + "path": "brass_acoustic_006-082-025.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22307 + }, + { + "path": "organ_electronic_028-031-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22308 + }, + { + "path": "keyboard_electronic_078-041-100.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22309 + }, + { + "path": "guitar_electronic_028-067-127.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22310 + }, + { + "path": "keyboard_electronic_098-043-127.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22311 + }, + { + "path": "organ_electronic_007-066-075.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22312 + }, + { + "path": "keyboard_synthetic_000-093-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22313 + }, + { + "path": "guitar_acoustic_014-105-075.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22314 + }, + { + "path": "guitar_electronic_028-097-127.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22315 + }, + { + "path": "organ_electronic_001-077-075.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22316 + }, + { + "path": "organ_electronic_104-083-100.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22317 + }, + { + "path": "guitar_electronic_028-060-100.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22318 + }, + { + "path": "keyboard_electronic_078-038-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22319 + }, + { + "path": "bass_electronic_018-052-127.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22320 + }, + { + "path": "guitar_electronic_028-095-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22321 + }, + { + "path": "vocal_synthetic_003-044-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22322 + }, + { + "path": "brass_acoustic_015-046-025.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22323 + }, + { + "path": "reed_acoustic_037-058-050.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22324 + }, + { + "path": "bass_synthetic_009-070-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22325 + }, + { + "path": "reed_acoustic_023-076-127.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22326 + }, + { + "path": "string_acoustic_057-088-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22327 + }, + { + "path": "bass_electronic_025-063-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22328 + }, + { + "path": "keyboard_electronic_003-023-025.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22329 + }, + { + "path": "bass_synthetic_098-066-025.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22330 + }, + { + "path": "keyboard_electronic_001-042-025.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22331 + }, + { + "path": "brass_acoustic_016-081-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22332 + }, + { + "path": "mallet_acoustic_062-075-025.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22333 + }, + { + "path": "guitar_acoustic_014-053-100.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22334 + }, + { + "path": "bass_synthetic_009-077-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22335 + }, + { + "path": "keyboard_acoustic_004-041-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22336 + }, + { + "path": "bass_synthetic_009-082-050.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22337 + }, + { + "path": "bass_synthetic_098-041-127.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22338 + }, + { + "path": "string_acoustic_014-065-050.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22339 + }, + { + "path": "organ_electronic_104-050-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22340 + }, + { + "path": "string_acoustic_071-035-075.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22341 + }, + { + "path": "guitar_acoustic_010-093-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22342 + }, + { + "path": "guitar_acoustic_015-077-050.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22343 + }, + { + "path": "guitar_acoustic_015-071-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22344 + }, + { + "path": "bass_electronic_025-028-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22345 + }, + { + "path": "vocal_acoustic_000-054-050.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22346 + }, + { + "path": "organ_electronic_007-086-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22347 + }, + { + "path": "organ_electronic_113-045-127.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22348 + }, + { + "path": "reed_acoustic_018-065-075.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22349 + }, + { + "path": "keyboard_synthetic_000-026-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22350 + }, + { + "path": "mallet_acoustic_062-024-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22351 + }, + { + "path": "flute_synthetic_000-036-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22352 + }, + { + "path": "string_acoustic_071-029-025.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22353 + }, + { + "path": "bass_synthetic_135-046-127.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22354 + }, + { + "path": "bass_synthetic_009-067-050.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22355 + }, + { + "path": "reed_acoustic_011-035-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22356 + }, + { + "path": "guitar_acoustic_015-054-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22357 + }, + { + "path": "mallet_acoustic_062-079-025.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22358 + }, + { + "path": "guitar_acoustic_014-052-127.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22359 + }, + { + "path": "vocal_synthetic_003-054-050.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22360 + }, + { + "path": "organ_electronic_001-069-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22361 + }, + { + "path": "string_acoustic_056-040-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22362 + }, + { + "path": "bass_synthetic_068-069-100.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22363 + }, + { + "path": "organ_electronic_104-033-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22364 + }, + { + "path": "keyboard_electronic_098-044-025.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22365 + }, + { + "path": "organ_electronic_113-070-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22366 + }, + { + "path": "bass_synthetic_135-027-025.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22367 + }, + { + "path": "reed_acoustic_011-045-050.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22368 + }, + { + "path": "mallet_acoustic_062-039-075.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22369 + }, + { + "path": "bass_electronic_025-032-127.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22370 + }, + { + "path": "guitar_acoustic_014-102-025.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22371 + }, + { + "path": "keyboard_electronic_003-092-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22372 + }, + { + "path": "bass_synthetic_098-059-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22373 + }, + { + "path": "bass_synthetic_033-065-050.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22374 + }, + { + "path": "guitar_acoustic_014-082-075.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22375 + }, + { + "path": "guitar_electronic_022-049-127.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22376 + }, + { + "path": "guitar_acoustic_015-044-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22377 + }, + { + "path": "bass_synthetic_009-041-075.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22378 + }, + { + "path": "keyboard_synthetic_000-025-100.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22379 + }, + { + "path": "string_acoustic_056-070-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22380 + }, + { + "path": "string_acoustic_014-045-100.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22381 + }, + { + "path": "keyboard_electronic_001-032-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22382 + }, + { + "path": "keyboard_electronic_069-076-075.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22383 + }, + { + "path": "organ_electronic_001-087-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22384 + }, + { + "path": "flute_acoustic_002-103-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22385 + }, + { + "path": "keyboard_electronic_078-045-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22386 + }, + { + "path": "brass_acoustic_046-057-050.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22387 + }, + { + "path": "organ_electronic_104-082-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22388 + }, + { + "path": "guitar_acoustic_015-074-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22389 + }, + { + "path": "organ_electronic_007-011-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22390 + }, + { + "path": "keyboard_electronic_003-073-050.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22391 + }, + { + "path": "keyboard_electronic_002-076-075.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22392 + }, + { + "path": "guitar_acoustic_015-098-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22393 + }, + { + "path": "brass_acoustic_016-085-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22394 + }, + { + "path": "reed_acoustic_023-033-075.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22395 + }, + { + "path": "bass_synthetic_009-081-100.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22396 + }, + { + "path": "keyboard_electronic_069-061-075.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22397 + }, + { + "path": "mallet_acoustic_062-104-050.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22398 + }, + { + "path": "keyboard_electronic_003-089-075.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22399 + }, + { + "path": "mallet_acoustic_047-065-050.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22400 + }, + { + "path": "string_acoustic_080-053-075.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22401 + }, + { + "path": "bass_synthetic_134-085-050.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22402 + }, + { + "path": "reed_acoustic_023-052-025.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22403 + }, + { + "path": "brass_acoustic_016-067-075.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22404 + }, + { + "path": "mallet_acoustic_056-044-025.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22405 + }, + { + "path": "vocal_synthetic_003-049-100.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22406 + }, + { + "path": "organ_electronic_104-075-025.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22407 + }, + { + "path": "guitar_acoustic_021-035-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22408 + }, + { + "path": "guitar_electronic_022-042-075.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22409 + }, + { + "path": "guitar_acoustic_015-058-100.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22410 + }, + { + "path": "guitar_electronic_022-078-050.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22411 + }, + { + "path": "organ_electronic_113-083-075.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22412 + }, + { + "path": "bass_synthetic_098-090-050.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22413 + }, + { + "path": "string_acoustic_071-038-025.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22414 + }, + { + "path": "keyboard_acoustic_004-084-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22415 + }, + { + "path": "keyboard_electronic_001-092-050.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22416 + }, + { + "path": "guitar_electronic_028-029-075.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22417 + }, + { + "path": "guitar_acoustic_021-025-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22418 + }, + { + "path": "keyboard_synthetic_000-064-127.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22419 + }, + { + "path": "keyboard_electronic_003-038-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22420 + }, + { + "path": "bass_electronic_025-038-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22421 + }, + { + "path": "keyboard_electronic_002-059-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22422 + }, + { + "path": "bass_synthetic_068-077-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22423 + }, + { + "path": "bass_synthetic_068-029-075.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22424 + }, + { + "path": "vocal_synthetic_003-027-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22425 + }, + { + "path": "guitar_electronic_028-030-075.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22426 + }, + { + "path": "string_acoustic_057-067-100.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22427 + }, + { + "path": "keyboard_electronic_003-025-075.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22428 + }, + { + "path": "reed_acoustic_037-068-050.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22429 + }, + { + "path": "keyboard_electronic_001-043-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22430 + }, + { + "path": "guitar_acoustic_014-069-127.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22431 + }, + { + "path": "keyboard_acoustic_004-096-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22432 + }, + { + "path": "mallet_acoustic_062-083-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22433 + }, + { + "path": "guitar_acoustic_030-077-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22434 + }, + { + "path": "organ_electronic_001-061-100.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22435 + }, + { + "path": "bass_synthetic_134-036-075.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22436 + }, + { + "path": "mallet_acoustic_062-098-025.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22437 + }, + { + "path": "guitar_acoustic_010-070-025.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22438 + }, + { + "path": "flute_synthetic_000-105-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22439 + }, + { + "path": "organ_electronic_028-066-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22440 + }, + { + "path": "keyboard_electronic_003-056-127.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22441 + }, + { + "path": "keyboard_electronic_001-060-050.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22442 + }, + { + "path": "brass_acoustic_006-068-127.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22443 + }, + { + "path": "string_acoustic_012-026-050.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22444 + }, + { + "path": "brass_acoustic_006-073-100.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22445 + }, + { + "path": "flute_synthetic_000-077-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22446 + }, + { + "path": "vocal_synthetic_003-101-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22447 + }, + { + "path": "bass_synthetic_134-052-050.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22448 + }, + { + "path": "organ_electronic_104-045-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22449 + }, + { + "path": "keyboard_electronic_003-061-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22450 + }, + { + "path": "bass_synthetic_134-045-100.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22451 + }, + { + "path": "keyboard_electronic_001-052-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22452 + }, + { + "path": "vocal_synthetic_003-106-100.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22453 + }, + { + "path": "keyboard_electronic_098-089-025.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22454 + }, + { + "path": "bass_synthetic_033-101-050.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22455 + }, + { + "path": "bass_synthetic_134-074-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22456 + }, + { + "path": "organ_electronic_028-031-100.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22457 + }, + { + "path": "guitar_acoustic_010-042-050.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22458 + }, + { + "path": "brass_acoustic_016-066-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22459 + }, + { + "path": "organ_electronic_113-051-127.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22460 + }, + { + "path": "mallet_acoustic_062-057-025.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22461 + }, + { + "path": "bass_synthetic_098-028-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22462 + }, + { + "path": "mallet_acoustic_056-113-075.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22463 + }, + { + "path": "brass_acoustic_059-062-025.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22464 + }, + { + "path": "keyboard_synthetic_000-070-127.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22465 + }, + { + "path": "keyboard_electronic_098-098-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22466 + }, + { + "path": "string_acoustic_014-034-075.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22467 + }, + { + "path": "organ_electronic_028-066-127.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22468 + }, + { + "path": "guitar_acoustic_010-074-100.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22469 + }, + { + "path": "guitar_electronic_028-032-050.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22470 + }, + { + "path": "guitar_acoustic_010-030-127.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22471 + }, + { + "path": "organ_electronic_113-075-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22472 + }, + { + "path": "organ_electronic_007-043-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22473 + }, + { + "path": "guitar_acoustic_030-107-050.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22474 + }, + { + "path": "organ_electronic_104-047-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22475 + }, + { + "path": "string_acoustic_056-059-075.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22476 + }, + { + "path": "guitar_acoustic_021-069-100.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22477 + }, + { + "path": "organ_electronic_007-076-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22478 + }, + { + "path": "bass_synthetic_134-075-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22479 + }, + { + "path": "mallet_acoustic_047-105-075.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22480 + }, + { + "path": "keyboard_electronic_001-054-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22481 + }, + { + "path": "bass_synthetic_068-075-025.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22482 + }, + { + "path": "keyboard_electronic_001-086-075.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22483 + }, + { + "path": "keyboard_electronic_003-047-127.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22484 + }, + { + "path": "keyboard_electronic_069-066-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22485 + }, + { + "path": "flute_acoustic_002-070-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22486 + }, + { + "path": "bass_synthetic_034-050-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22487 + }, + { + "path": "guitar_acoustic_014-086-050.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22488 + }, + { + "path": "keyboard_electronic_002-063-050.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22489 + }, + { + "path": "keyboard_electronic_078-046-127.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22490 + }, + { + "path": "bass_electronic_018-023-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22491 + }, + { + "path": "mallet_acoustic_056-075-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22492 + }, + { + "path": "keyboard_electronic_003-028-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22493 + }, + { + "path": "keyboard_electronic_069-059-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22494 + }, + { + "path": "keyboard_electronic_001-104-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22495 + }, + { + "path": "keyboard_synthetic_000-094-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22496 + }, + { + "path": "organ_electronic_113-071-100.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22497 + }, + { + "path": "guitar_acoustic_021-095-075.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22498 + }, + { + "path": "guitar_acoustic_015-067-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22499 + }, + { + "path": "keyboard_electronic_001-031-050.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22500 + }, + { + "path": "bass_synthetic_098-070-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22501 + }, + { + "path": "guitar_acoustic_015-059-050.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22502 + }, + { + "path": "bass_synthetic_034-023-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22503 + }, + { + "path": "bass_synthetic_135-049-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22504 + }, + { + "path": "keyboard_synthetic_000-024-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22505 + }, + { + "path": "string_acoustic_056-045-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22506 + }, + { + "path": "guitar_acoustic_010-035-025.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22507 + }, + { + "path": "guitar_acoustic_015-069-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22508 + }, + { + "path": "brass_acoustic_015-076-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22509 + }, + { + "path": "guitar_electronic_022-047-100.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22510 + }, + { + "path": "reed_acoustic_023-065-050.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22511 + }, + { + "path": "brass_acoustic_006-027-050.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22512 + }, + { + "path": "flute_acoustic_002-084-050.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22513 + }, + { + "path": "brass_acoustic_006-026-127.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22514 + }, + { + "path": "guitar_acoustic_015-030-075.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22515 + }, + { + "path": "keyboard_electronic_069-064-100.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22516 + }, + { + "path": "bass_synthetic_134-080-075.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22517 + }, + { + "path": "guitar_acoustic_015-104-100.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22518 + }, + { + "path": "organ_electronic_113-075-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22519 + }, + { + "path": "guitar_acoustic_014-108-050.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22520 + }, + { + "path": "organ_electronic_007-041-100.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22521 + }, + { + "path": "reed_acoustic_018-079-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22522 + }, + { + "path": "keyboard_synthetic_000-076-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22523 + }, + { + "path": "keyboard_electronic_078-083-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22524 + }, + { + "path": "bass_synthetic_009-080-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22525 + }, + { + "path": "keyboard_electronic_098-049-127.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22526 + }, + { + "path": "guitar_acoustic_014-072-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22527 + }, + { + "path": "keyboard_electronic_098-042-025.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22528 + }, + { + "path": "guitar_acoustic_015-048-100.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22529 + }, + { + "path": "bass_synthetic_134-081-025.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22530 + }, + { + "path": "vocal_synthetic_003-084-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22531 + }, + { + "path": "organ_electronic_007-070-075.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22532 + }, + { + "path": "reed_acoustic_018-061-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22533 + }, + { + "path": "organ_electronic_007-080-050.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22534 + }, + { + "path": "keyboard_electronic_002-083-075.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22535 + }, + { + "path": "keyboard_electronic_001-059-075.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22536 + }, + { + "path": "keyboard_electronic_003-082-075.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22537 + }, + { + "path": "guitar_acoustic_014-039-050.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22538 + }, + { + "path": "reed_acoustic_018-083-025.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22539 + }, + { + "path": "vocal_synthetic_003-061-025.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22540 + }, + { + "path": "bass_electronic_025-035-127.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22541 + }, + { + "path": "brass_acoustic_059-031-050.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22542 + }, + { + "path": "keyboard_synthetic_000-040-050.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22543 + }, + { + "path": "keyboard_electronic_003-023-100.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22544 + }, + { + "path": "keyboard_acoustic_004-091-127.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22545 + }, + { + "path": "string_acoustic_057-056-127.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22546 + }, + { + "path": "brass_acoustic_046-095-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22547 + }, + { + "path": "reed_acoustic_037-035-050.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22548 + }, + { + "path": "bass_synthetic_033-085-050.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22549 + }, + { + "path": "guitar_acoustic_014-027-127.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22550 + }, + { + "path": "bass_synthetic_135-055-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22551 + }, + { + "path": "organ_electronic_104-067-127.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22552 + }, + { + "path": "organ_electronic_007-092-127.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22553 + }, + { + "path": "vocal_synthetic_003-039-025.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22554 + }, + { + "path": "brass_acoustic_016-061-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22555 + }, + { + "path": "flute_acoustic_002-095-100.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22556 + }, + { + "path": "string_acoustic_071-049-025.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22557 + }, + { + "path": "guitar_acoustic_030-079-127.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22558 + }, + { + "path": "keyboard_electronic_069-084-100.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22559 + }, + { + "path": "bass_synthetic_068-046-075.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22560 + }, + { + "path": "keyboard_synthetic_000-062-100.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22561 + }, + { + "path": "reed_acoustic_011-067-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22562 + }, + { + "path": "keyboard_synthetic_000-034-050.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22563 + }, + { + "path": "organ_electronic_104-039-025.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22564 + }, + { + "path": "bass_synthetic_009-013-127.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22565 + }, + { + "path": "keyboard_electronic_098-043-075.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22566 + }, + { + "path": "guitar_acoustic_015-094-127.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22567 + }, + { + "path": "bass_synthetic_033-064-075.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22568 + }, + { + "path": "reed_acoustic_023-086-075.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22569 + }, + { + "path": "reed_acoustic_011-068-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22570 + }, + { + "path": "guitar_acoustic_014-070-025.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22571 + }, + { + "path": "keyboard_electronic_003-087-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22572 + }, + { + "path": "mallet_acoustic_062-084-100.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22573 + }, + { + "path": "organ_electronic_007-028-025.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22574 + }, + { + "path": "brass_acoustic_046-071-100.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22575 + }, + { + "path": "keyboard_electronic_069-048-075.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22576 + }, + { + "path": "keyboard_electronic_069-082-100.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22577 + }, + { + "path": "guitar_acoustic_010-057-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22578 + }, + { + "path": "keyboard_electronic_078-069-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22579 + }, + { + "path": "bass_synthetic_033-030-025.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22580 + }, + { + "path": "mallet_acoustic_047-103-075.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22581 + }, + { + "path": "organ_electronic_028-097-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22582 + }, + { + "path": "guitar_acoustic_010-093-100.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22583 + }, + { + "path": "string_acoustic_057-076-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22584 + }, + { + "path": "keyboard_electronic_002-024-050.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22585 + }, + { + "path": "string_acoustic_056-066-127.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22586 + }, + { + "path": "bass_synthetic_033-072-075.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22587 + }, + { + "path": "bass_synthetic_033-069-100.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22588 + }, + { + "path": "guitar_acoustic_021-040-127.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22589 + }, + { + "path": "bass_synthetic_009-036-025.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22590 + }, + { + "path": "keyboard_electronic_002-041-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22591 + }, + { + "path": "organ_electronic_057-034-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22592 + }, + { + "path": "bass_synthetic_009-052-100.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22593 + }, + { + "path": "bass_synthetic_134-098-100.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22594 + }, + { + "path": "organ_electronic_057-027-075.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22595 + }, + { + "path": "guitar_acoustic_010-039-127.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22596 + }, + { + "path": "flute_acoustic_002-092-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22597 + }, + { + "path": "string_acoustic_014-028-127.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22598 + }, + { + "path": "keyboard_electronic_069-050-025.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22599 + }, + { + "path": "keyboard_electronic_001-065-025.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22600 + }, + { + "path": "guitar_electronic_022-036-100.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22601 + }, + { + "path": "guitar_acoustic_014-100-075.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22602 + }, + { + "path": "bass_synthetic_098-061-050.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22603 + }, + { + "path": "guitar_electronic_022-038-050.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22604 + }, + { + "path": "bass_synthetic_033-099-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22605 + }, + { + "path": "keyboard_electronic_078-060-050.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22606 + }, + { + "path": "guitar_acoustic_015-082-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22607 + }, + { + "path": "guitar_acoustic_014-101-050.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22608 + }, + { + "path": "bass_electronic_018-027-075.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22609 + }, + { + "path": "bass_electronic_018-061-127.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22610 + }, + { + "path": "organ_electronic_113-053-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22611 + }, + { + "path": "bass_synthetic_135-068-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22612 + }, + { + "path": "string_acoustic_056-038-050.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22613 + }, + { + "path": "bass_synthetic_134-085-127.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22614 + }, + { + "path": "keyboard_electronic_001-057-127.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22615 + }, + { + "path": "brass_acoustic_006-077-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22616 + }, + { + "path": "string_acoustic_014-031-050.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22617 + }, + { + "path": "guitar_acoustic_015-103-127.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22618 + }, + { + "path": "reed_acoustic_037-041-075.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22619 + }, + { + "path": "bass_synthetic_068-098-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22620 + }, + { + "path": "bass_electronic_018-060-127.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22621 + }, + { + "path": "keyboard_electronic_003-053-100.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22622 + }, + { + "path": "organ_electronic_104-055-075.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22623 + }, + { + "path": "bass_synthetic_134-100-100.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22624 + }, + { + "path": "mallet_acoustic_062-050-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22625 + }, + { + "path": "keyboard_electronic_002-100-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22626 + }, + { + "path": "bass_synthetic_098-071-050.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22627 + }, + { + "path": "guitar_acoustic_010-084-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22628 + }, + { + "path": "string_acoustic_057-069-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22629 + }, + { + "path": "organ_electronic_104-026-127.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22630 + }, + { + "path": "brass_acoustic_006-034-075.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22631 + }, + { + "path": "bass_synthetic_009-058-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22632 + }, + { + "path": "guitar_acoustic_014-096-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22633 + }, + { + "path": "organ_electronic_104-077-025.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22634 + }, + { + "path": "string_acoustic_057-059-025.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22635 + }, + { + "path": "bass_synthetic_034-093-075.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22636 + }, + { + "path": "string_acoustic_014-063-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22637 + }, + { + "path": "keyboard_electronic_078-061-127.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22638 + }, + { + "path": "bass_synthetic_034-085-075.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22639 + }, + { + "path": "reed_acoustic_018-067-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22640 + }, + { + "path": "bass_electronic_018-057-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22641 + }, + { + "path": "bass_synthetic_098-038-100.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22642 + }, + { + "path": "keyboard_electronic_003-071-100.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22643 + }, + { + "path": "organ_electronic_113-053-050.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22644 + }, + { + "path": "keyboard_electronic_002-089-127.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22645 + }, + { + "path": "organ_electronic_104-053-025.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22646 + }, + { + "path": "keyboard_acoustic_004-079-075.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22647 + }, + { + "path": "organ_electronic_028-042-050.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22648 + }, + { + "path": "guitar_acoustic_015-027-100.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22649 + }, + { + "path": "guitar_electronic_022-086-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22650 + }, + { + "path": "organ_electronic_007-096-075.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22651 + }, + { + "path": "keyboard_electronic_003-086-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22652 + }, + { + "path": "mallet_acoustic_047-095-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22653 + }, + { + "path": "bass_electronic_018-036-025.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22654 + }, + { + "path": "guitar_acoustic_010-021-127.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22655 + }, + { + "path": "reed_acoustic_011-074-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22656 + }, + { + "path": "guitar_electronic_028-079-127.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22657 + }, + { + "path": "bass_synthetic_034-085-025.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22658 + }, + { + "path": "organ_electronic_007-010-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22659 + }, + { + "path": "string_acoustic_012-055-075.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22660 + }, + { + "path": "guitar_acoustic_010-032-075.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22661 + }, + { + "path": "guitar_acoustic_021-090-075.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22662 + }, + { + "path": "bass_synthetic_033-056-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22663 + }, + { + "path": "guitar_electronic_022-073-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22664 + }, + { + "path": "guitar_acoustic_010-070-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22665 + }, + { + "path": "keyboard_acoustic_004-031-075.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22666 + }, + { + "path": "brass_acoustic_006-056-127.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22667 + }, + { + "path": "keyboard_electronic_098-074-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22668 + }, + { + "path": "guitar_acoustic_010-085-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22669 + }, + { + "path": "string_acoustic_012-050-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22670 + }, + { + "path": "keyboard_electronic_002-025-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22671 + }, + { + "path": "brass_acoustic_046-081-050.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22672 + }, + { + "path": "keyboard_electronic_001-045-025.wav", + "question": "The MIDI velocity for this note falls at what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22673 + }, + { + "path": "keyboard_acoustic_004-091-100.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22674 + }, + { + "path": "guitar_acoustic_015-021-075.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22675 + }, + { + "path": "keyboard_electronic_098-103-127.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22676 + }, + { + "path": "bass_synthetic_034-063-025.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22677 + }, + { + "path": "bass_synthetic_134-034-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22678 + }, + { + "path": "brass_acoustic_059-052-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22679 + }, + { + "path": "brass_acoustic_015-066-025.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22680 + }, + { + "path": "bass_synthetic_098-061-127.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22681 + }, + { + "path": "organ_electronic_028-034-100.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22682 + }, + { + "path": "keyboard_electronic_001-077-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22683 + }, + { + "path": "brass_acoustic_059-048-050.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22684 + }, + { + "path": "keyboard_electronic_002-090-050.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22685 + }, + { + "path": "string_acoustic_014-050-075.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22686 + }, + { + "path": "bass_synthetic_135-087-075.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22687 + }, + { + "path": "bass_synthetic_009-019-127.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22688 + }, + { + "path": "keyboard_synthetic_000-098-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22689 + }, + { + "path": "mallet_acoustic_047-066-100.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22690 + }, + { + "path": "keyboard_electronic_069-039-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22691 + }, + { + "path": "vocal_synthetic_003-065-075.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22692 + }, + { + "path": "brass_acoustic_015-058-050.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22693 + }, + { + "path": "reed_acoustic_023-041-127.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22694 + }, + { + "path": "flute_synthetic_000-037-127.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22695 + }, + { + "path": "guitar_acoustic_010-024-050.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22696 + }, + { + "path": "string_acoustic_056-044-025.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22697 + }, + { + "path": "bass_electronic_018-023-050.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22698 + }, + { + "path": "keyboard_electronic_002-067-127.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22699 + }, + { + "path": "guitar_electronic_028-073-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22700 + }, + { + "path": "bass_synthetic_033-041-050.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22701 + }, + { + "path": "guitar_acoustic_021-092-127.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22702 + }, + { + "path": "vocal_synthetic_003-037-075.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22703 + }, + { + "path": "reed_acoustic_011-043-075.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22704 + }, + { + "path": "guitar_electronic_028-093-127.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22705 + }, + { + "path": "organ_electronic_028-084-075.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22706 + }, + { + "path": "guitar_acoustic_021-046-127.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22707 + }, + { + "path": "keyboard_acoustic_004-061-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22708 + }, + { + "path": "keyboard_electronic_069-075-025.wav", + "question": "Select the MIDI velocity that represents this note.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22709 + }, + { + "path": "keyboard_acoustic_004-075-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22710 + }, + { + "path": "keyboard_electronic_098-044-075.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22711 + }, + { + "path": "keyboard_electronic_002-046-127.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22712 + }, + { + "path": "guitar_acoustic_021-048-025.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22713 + }, + { + "path": "organ_electronic_001-084-050.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22714 + }, + { + "path": "vocal_synthetic_003-054-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22715 + }, + { + "path": "keyboard_electronic_098-045-075.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22716 + }, + { + "path": "organ_electronic_104-029-025.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22717 + }, + { + "path": "organ_electronic_001-076-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22718 + }, + { + "path": "keyboard_synthetic_000-033-025.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22719 + }, + { + "path": "brass_acoustic_006-057-100.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22720 + }, + { + "path": "keyboard_acoustic_004-065-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22721 + }, + { + "path": "bass_synthetic_135-094-100.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22722 + }, + { + "path": "guitar_acoustic_015-108-127.wav", + "question": "Determine the MIDI velocity for the note in question.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22723 + }, + { + "path": "bass_synthetic_134-075-075.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22724 + }, + { + "path": "reed_acoustic_011-054-025.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22725 + }, + { + "path": "brass_acoustic_015-054-127.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22726 + }, + { + "path": "brass_acoustic_046-104-025.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22727 + }, + { + "path": "reed_acoustic_018-084-127.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22728 + }, + { + "path": "organ_electronic_007-082-100.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22729 + }, + { + "path": "bass_synthetic_033-045-075.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22730 + }, + { + "path": "guitar_acoustic_030-064-025.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22731 + }, + { + "path": "bass_synthetic_033-038-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22732 + }, + { + "path": "bass_electronic_018-049-075.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22733 + }, + { + "path": "keyboard_synthetic_000-099-050.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22734 + }, + { + "path": "string_acoustic_014-026-127.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22735 + }, + { + "path": "organ_electronic_007-080-100.wav", + "question": "The MIDI velocity designation for this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22736 + }, + { + "path": "bass_synthetic_135-083-127.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22737 + }, + { + "path": "mallet_acoustic_047-068-050.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22738 + }, + { + "path": "string_acoustic_057-062-075.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22739 + }, + { + "path": "string_acoustic_057-081-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22740 + }, + { + "path": "bass_synthetic_009-047-100.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22741 + }, + { + "path": "bass_synthetic_009-094-100.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22742 + }, + { + "path": "organ_electronic_057-052-100.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22743 + }, + { + "path": "guitar_electronic_022-085-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22744 + }, + { + "path": "keyboard_electronic_002-101-075.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22745 + }, + { + "path": "bass_synthetic_068-046-025.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22746 + }, + { + "path": "keyboard_electronic_002-055-050.wav", + "question": "What numerical value is the MIDI velocity for this note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22747 + }, + { + "path": "reed_acoustic_018-070-100.wav", + "question": "What MIDI velocity does this note translate to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22748 + }, + { + "path": "organ_electronic_113-056-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22749 + }, + { + "path": "guitar_acoustic_010-088-100.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22750 + }, + { + "path": "bass_electronic_027-046-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22751 + }, + { + "path": "brass_acoustic_046-084-075.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22752 + }, + { + "path": "bass_electronic_027-029-075.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22753 + }, + { + "path": "organ_electronic_028-028-100.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22754 + }, + { + "path": "mallet_acoustic_047-082-050.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22755 + }, + { + "path": "bass_synthetic_134-052-075.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22756 + }, + { + "path": "bass_synthetic_033-033-025.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22757 + }, + { + "path": "bass_synthetic_034-046-025.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22758 + }, + { + "path": "bass_electronic_018-041-100.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22759 + }, + { + "path": "keyboard_electronic_001-049-127.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22760 + }, + { + "path": "bass_synthetic_134-044-025.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22761 + }, + { + "path": "organ_electronic_104-064-075.wav", + "question": "The MIDI velocity for this note can be identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22762 + }, + { + "path": "keyboard_electronic_078-041-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22763 + }, + { + "path": "bass_synthetic_068-084-050.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22764 + }, + { + "path": "organ_electronic_104-034-050.wav", + "question": "The note's loudness equates to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22765 + }, + { + "path": "reed_acoustic_011-062-050.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22766 + }, + { + "path": "guitar_acoustic_030-062-127.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22767 + }, + { + "path": "reed_acoustic_037-034-127.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22768 + }, + { + "path": "guitar_acoustic_021-077-100.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22769 + }, + { + "path": "string_acoustic_057-082-050.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22770 + }, + { + "path": "bass_synthetic_098-060-100.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22771 + }, + { + "path": "string_acoustic_056-073-100.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22772 + }, + { + "path": "bass_synthetic_009-021-050.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22773 + }, + { + "path": "guitar_acoustic_021-047-075.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22774 + }, + { + "path": "keyboard_acoustic_004-051-050.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22775 + }, + { + "path": "bass_synthetic_135-082-127.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22776 + }, + { + "path": "keyboard_electronic_002-027-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22777 + }, + { + "path": "keyboard_synthetic_000-039-075.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22778 + }, + { + "path": "flute_synthetic_000-071-075.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22779 + }, + { + "path": "bass_synthetic_009-091-050.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22780 + }, + { + "path": "keyboard_electronic_001-031-025.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22781 + }, + { + "path": "reed_acoustic_023-095-025.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22782 + }, + { + "path": "bass_electronic_018-039-025.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22783 + }, + { + "path": "string_acoustic_014-025-075.wav", + "question": "This note has a MIDI velocity of what number?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22784 + }, + { + "path": "organ_electronic_113-081-100.wav", + "question": "What is the MIDI velocity value of this isolated note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22785 + }, + { + "path": "keyboard_electronic_002-074-075.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22786 + }, + { + "path": "keyboard_synthetic_000-082-025.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22787 + }, + { + "path": "bass_synthetic_135-044-050.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22788 + }, + { + "path": "brass_acoustic_006-087-050.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22789 + }, + { + "path": "bass_synthetic_034-045-025.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22790 + }, + { + "path": "flute_synthetic_000-074-050.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22791 + }, + { + "path": "guitar_acoustic_021-099-050.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22792 + }, + { + "path": "string_acoustic_012-029-127.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22793 + }, + { + "path": "bass_synthetic_134-065-127.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22794 + }, + { + "path": "organ_electronic_001-075-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22795 + }, + { + "path": "guitar_acoustic_010-100-100.wav", + "question": "The MIDI velocity for the lone note in this music is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22796 + }, + { + "path": "organ_electronic_028-042-025.wav", + "question": "The MIDI velocity number for the note heard is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22797 + }, + { + "path": "organ_electronic_007-078-050.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22798 + }, + { + "path": "keyboard_electronic_002-075-075.wav", + "question": "The MIDI velocity of the played note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22799 + }, + { + "path": "mallet_acoustic_047-098-127.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22800 + }, + { + "path": "flute_synthetic_000-057-100.wav", + "question": "What MIDI velocity is the note equivalent to?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22801 + }, + { + "path": "organ_electronic_001-059-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22802 + }, + { + "path": "bass_synthetic_033-063-075.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22803 + }, + { + "path": "guitar_electronic_022-077-050.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22804 + }, + { + "path": "bass_electronic_018-029-050.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22805 + }, + { + "path": "guitar_acoustic_015-106-127.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22806 + }, + { + "path": "guitar_acoustic_021-086-127.wav", + "question": "The note in this music piece has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22807 + }, + { + "path": "organ_electronic_028-071-127.wav", + "question": "Select the MIDI velocity that aligns with this note's volume.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22808 + }, + { + "path": "guitar_acoustic_021-061-127.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22809 + }, + { + "path": "bass_synthetic_009-015-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22810 + }, + { + "path": "reed_acoustic_011-041-075.wav", + "question": "The MIDI velocity of this note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22811 + }, + { + "path": "guitar_acoustic_021-072-075.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22812 + }, + { + "path": "bass_synthetic_034-026-050.wav", + "question": "What is the MIDI velocity level of the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22813 + }, + { + "path": "vocal_synthetic_003-061-127.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22814 + }, + { + "path": "bass_synthetic_098-104-050.wav", + "question": "Which MIDI velocity corresponds to the note played?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22815 + }, + { + "path": "organ_electronic_057-062-100.wav", + "question": "The MIDI velocity of the note being heard is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22816 + }, + { + "path": "bass_synthetic_034-021-050.wav", + "question": "The precise MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22817 + }, + { + "path": "organ_electronic_028-093-075.wav", + "question": "What is the correct MIDI velocity of the note you're hearing?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22818 + }, + { + "path": "mallet_acoustic_047-103-025.wav", + "question": "The MIDI velocity associated with this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22819 + }, + { + "path": "reed_acoustic_011-037-050.wav", + "question": "This musical note's MIDI velocity is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22820 + }, + { + "path": "organ_electronic_113-043-100.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22821 + }, + { + "path": "keyboard_acoustic_004-045-100.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22822 + }, + { + "path": "brass_acoustic_046-052-075.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22823 + }, + { + "path": "mallet_acoustic_062-042-075.wav", + "question": "The MIDI velocity value for this music note is what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22824 + }, + { + "path": "keyboard_acoustic_004-098-127.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22825 + }, + { + "path": "vocal_acoustic_000-068-075.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22826 + }, + { + "path": "guitar_acoustic_021-102-100.wav", + "question": "This note corresponds to what MIDI velocity value?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22827 + }, + { + "path": "reed_acoustic_023-080-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22828 + }, + { + "path": "organ_electronic_104-075-100.wav", + "question": "What is the MIDI velocity of the note in this music?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22829 + }, + { + "path": "mallet_acoustic_056-043-050.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22830 + }, + { + "path": "guitar_acoustic_010-025-100.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22831 + }, + { + "path": "bass_synthetic_034-034-025.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22832 + }, + { + "path": "organ_electronic_104-071-025.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22833 + }, + { + "path": "organ_electronic_028-054-127.wav", + "question": "This note's volume translates to which MIDI velocity?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22834 + }, + { + "path": "keyboard_synthetic_000-023-075.wav", + "question": "The MIDI velocity for this musical tone is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22835 + }, + { + "path": "mallet_acoustic_047-088-025.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22836 + }, + { + "path": "bass_synthetic_068-030-075.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22837 + }, + { + "path": "guitar_acoustic_015-043-127.wav", + "question": "Identify the MIDI velocity that matches this note's dynamics.", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22838 + }, + { + "path": "string_acoustic_056-041-127.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22839 + }, + { + "path": "reed_acoustic_037-042-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22840 + }, + { + "path": "flute_synthetic_000-026-025.wav", + "question": "The MIDI velocity of this music note is what value?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22841 + }, + { + "path": "string_acoustic_080-032-075.wav", + "question": "The sound of this note corresponds to which MIDI velocity?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22842 + }, + { + "path": "brass_acoustic_016-080-100.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22843 + }, + { + "path": "guitar_acoustic_021-059-025.wav", + "question": "The MIDI velocity that denotes this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22844 + }, + { + "path": "keyboard_synthetic_000-046-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22845 + }, + { + "path": "bass_synthetic_033-027-050.wav", + "question": "Identify the MIDI velocity of the single musical note.", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22846 + }, + { + "path": "keyboard_electronic_069-077-025.wav", + "question": "The MIDI velocity of the musical note is identified as what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22847 + }, + { + "path": "bass_synthetic_009-035-025.wav", + "question": "What is the MIDI velocity number for this note's intensity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22848 + }, + { + "path": "string_acoustic_057-066-100.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22849 + }, + { + "path": "guitar_acoustic_010-031-100.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22850 + }, + { + "path": "bass_electronic_018-047-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22851 + }, + { + "path": "brass_acoustic_006-032-025.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22852 + }, + { + "path": "guitar_acoustic_014-061-100.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22853 + }, + { + "path": "bass_electronic_018-050-127.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22854 + }, + { + "path": "brass_acoustic_059-039-075.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22855 + }, + { + "path": "organ_electronic_028-029-025.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22856 + }, + { + "path": "guitar_acoustic_010-022-050.wav", + "question": "The single musical note has a MIDI velocity of what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22857 + }, + { + "path": "organ_electronic_001-078-025.wav", + "question": "What is the MIDI velocity assigned to this musical sound?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22858 + }, + { + "path": "bass_electronic_027-058-075.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22859 + }, + { + "path": "string_acoustic_080-028-025.wav", + "question": "The MIDI velocity for this note is determined as what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22860 + }, + { + "path": "bass_electronic_025-036-050.wav", + "question": "What MIDI velocity is being represented by this note?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_100", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22861 + }, + { + "path": "keyboard_electronic_002-094-025.wav", + "question": "This isolated note's MIDI velocity translates to what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_25", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_25", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22862 + }, + { + "path": "guitar_electronic_022-043-050.wav", + "question": "The specific MIDI velocity for this note is what?", + "choice_a": "midi_velocity_25", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_50", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22863 + }, + { + "path": "reed_acoustic_023-076-100.wav", + "question": "The note at this timestamp has a MIDI velocity of what?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22864 + }, + { + "path": "guitar_acoustic_010-053-075.wav", + "question": "What MIDI velocity does this note's loudness correspond to?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22865 + }, + { + "path": "keyboard_electronic_001-029-100.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_100", + "answer_gt": "midi_velocity_100", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22866 + }, + { + "path": "keyboard_acoustic_004-085-075.wav", + "question": "Identify the MIDI velocity that this note falls under.", + "choice_a": "midi_velocity_75", + "choice_b": "midi_velocity_50", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_127", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22867 + }, + { + "path": "guitar_electronic_022-088-127.wav", + "question": "This single note registers as which MIDI velocity?", + "choice_a": "midi_velocity_127", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_25", + "choice_d": "midi_velocity_50", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22868 + }, + { + "path": "bass_synthetic_134-057-075.wav", + "question": "The MIDI velocity of this distinct note is what?", + "choice_a": "midi_velocity_50", + "choice_b": "midi_velocity_100", + "choice_c": "midi_velocity_75", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22869 + }, + { + "path": "mallet_acoustic_062-072-075.wav", + "question": "The MIDI velocity corresponding to this note is what?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_127", + "choice_c": "midi_velocity_50", + "choice_d": "midi_velocity_75", + "answer_gt": "midi_velocity_75", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22870 + }, + { + "path": "bass_synthetic_068-104-127.wav", + "question": "What is the MIDI velocity of this note at the current moment?", + "choice_a": "midi_velocity_100", + "choice_b": "midi_velocity_75", + "choice_c": "midi_velocity_127", + "choice_d": "midi_velocity_25", + "answer_gt": "midi_velocity_127", + "task_name": "Music_Midi_Velocity_Analysis", + "dataset_name": "nsynth", + "uniq_id": 22871 + }, + { + "path": "00000108.wav", + "question": "Is the guzheng louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22872 + }, + { + "path": "00003649.wav", + "question": "Are there violin and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22873 + }, + { + "path": "00007595.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22874 + }, + { + "path": "00007503.wav", + "question": "Is there a banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22875 + }, + { + "path": "00003620.wav", + "question": "Is the pipa more rhythmic than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22876 + }, + { + "path": "00001688.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22877 + }, + { + "path": "00008276.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22878 + }, + { + "path": "00003987.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22879 + }, + { + "path": "00003501.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22880 + }, + { + "path": "00000666.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22881 + }, + { + "path": "00004625.wav", + "question": "Is the clarinet louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22882 + }, + { + "path": "00006190.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22883 + }, + { + "path": "00007742.wav", + "question": "Are there flute and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22884 + }, + { + "path": "00004160.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22885 + }, + { + "path": "00000434.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22886 + }, + { + "path": "00003206.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22887 + }, + { + "path": "00001554.wav", + "question": "Is there a guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22888 + }, + { + "path": "00002376.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22889 + }, + { + "path": "00006426.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22890 + }, + { + "path": "00005937.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22891 + }, + { + "path": "00004460.wav", + "question": "Are there cello and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22892 + }, + { + "path": "00000864.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22893 + }, + { + "path": "00000399.wav", + "question": "Is there a suona sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22894 + }, + { + "path": "00007855.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22895 + }, + { + "path": "00003220.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22896 + }, + { + "path": "00003254.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22897 + }, + { + "path": "00007772.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22898 + }, + { + "path": "00003893.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22899 + }, + { + "path": "00000269.wav", + "question": "Are there piano and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22900 + }, + { + "path": "00005183.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22901 + }, + { + "path": "00008099.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22902 + }, + { + "path": "00007662.wav", + "question": "Is there a bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22903 + }, + { + "path": "00006846.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22904 + }, + { + "path": "00006829.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22905 + }, + { + "path": "00007185.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22906 + }, + { + "path": "00001453.wav", + "question": "Are there trumpet and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22907 + }, + { + "path": "00008182.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22908 + }, + { + "path": "00006614.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22909 + }, + { + "path": "00004725.wav", + "question": "Are there suona and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22910 + }, + { + "path": "00004555.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22911 + }, + { + "path": "00002592.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22912 + }, + { + "path": "00008345.wav", + "question": "Are there trumpet and electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22913 + }, + { + "path": "00005459.wav", + "question": "Is the cello more rhythmic than the electric_bass?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22914 + }, + { + "path": "00000725.wav", + "question": "Is the pipa playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22915 + }, + { + "path": "00007886.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22916 + }, + { + "path": "00000801.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22917 + }, + { + "path": "00004884.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22918 + }, + { + "path": "00001972.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22919 + }, + { + "path": "00001723.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22920 + }, + { + "path": "00005595.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22921 + }, + { + "path": "00008771.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22922 + }, + { + "path": "00005493.wav", + "question": "Is there a electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22923 + }, + { + "path": "00006598.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22924 + }, + { + "path": "00007052.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22925 + }, + { + "path": "00000721.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22926 + }, + { + "path": "00003353.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22927 + }, + { + "path": "00001581.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22928 + }, + { + "path": "00001342.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22929 + }, + { + "path": "00008940.wav", + "question": "Is there a congas sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22930 + }, + { + "path": "00003029.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22931 + }, + { + "path": "00008467.wav", + "question": "Is the ukulele louder than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22932 + }, + { + "path": "00007533.wav", + "question": "Is there a banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22933 + }, + { + "path": "00007899.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22934 + }, + { + "path": "00002743.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22935 + }, + { + "path": "00003394.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22936 + }, + { + "path": "00000472.wav", + "question": "Is the bagpipe louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22937 + }, + { + "path": "00002126.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22938 + }, + { + "path": "00005324.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22939 + }, + { + "path": "00005959.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22940 + }, + { + "path": "00007867.wav", + "question": "Is there a suona sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22941 + }, + { + "path": "00007036.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22942 + }, + { + "path": "00000334.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22943 + }, + { + "path": "00005011.wav", + "question": "Are there violin and trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22944 + }, + { + "path": "00006552.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22945 + }, + { + "path": "00005699.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22946 + }, + { + "path": "00001598.wav", + "question": "Is there a banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22947 + }, + { + "path": "00003796.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22948 + }, + { + "path": "00005352.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22949 + }, + { + "path": "00008383.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22950 + }, + { + "path": "00004178.wav", + "question": "Are there piano and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22951 + }, + { + "path": "00008434.wav", + "question": "Are there acoustic_guitar and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22952 + }, + { + "path": "00001213.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22953 + }, + { + "path": "00000129.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22954 + }, + { + "path": "00000129.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22955 + }, + { + "path": "00008014.wav", + "question": "Are there acoustic_guitar and erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22956 + }, + { + "path": "00005712.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22957 + }, + { + "path": "00007395.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22958 + }, + { + "path": "00001020.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22959 + }, + { + "path": "00002875.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22960 + }, + { + "path": "00005091.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22961 + }, + { + "path": "00005613.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22962 + }, + { + "path": "00002603.wav", + "question": "Are there trumpet and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22963 + }, + { + "path": "00003873.wav", + "question": "Are there flute and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22964 + }, + { + "path": "00004598.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22965 + }, + { + "path": "00004509.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22966 + }, + { + "path": "00001950.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22967 + }, + { + "path": "00008162.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22968 + }, + { + "path": "00001976.wav", + "question": "Is the electric_bass playing longer than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22969 + }, + { + "path": "00004257.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22970 + }, + { + "path": "00002469.wav", + "question": "Are there violin and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22971 + }, + { + "path": "00000098.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22972 + }, + { + "path": "00006644.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22973 + }, + { + "path": "00008737.wav", + "question": "Is there a guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22974 + }, + { + "path": "00001824.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22975 + }, + { + "path": "00003576.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22976 + }, + { + "path": "00002413.wav", + "question": "Are there acoustic_guitar and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22977 + }, + { + "path": "00000768.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22978 + }, + { + "path": "00002958.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22979 + }, + { + "path": "00007006.wav", + "question": "Is the piano louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22980 + }, + { + "path": "00007141.wav", + "question": "Are there clarinet and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22981 + }, + { + "path": "00006889.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22982 + }, + { + "path": "00000590.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22983 + }, + { + "path": "00008296.wav", + "question": "Are there erhu and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22984 + }, + { + "path": "00004686.wav", + "question": "Are there bassoon and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22985 + }, + { + "path": "00001475.wav", + "question": "Is there a bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22986 + }, + { + "path": "00006869.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22987 + }, + { + "path": "00001494.wav", + "question": "Are there violin and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22988 + }, + { + "path": "00002839.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22989 + }, + { + "path": "00005474.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22990 + }, + { + "path": "00002976.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22991 + }, + { + "path": "00007685.wav", + "question": "Is the violin more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22992 + }, + { + "path": "00005667.wav", + "question": "Is the piano more rhythmic than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22993 + }, + { + "path": "00003079.wav", + "question": "Are there piano and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22994 + }, + { + "path": "00007206.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22995 + }, + { + "path": "00006837.wav", + "question": "Is the erhu louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22996 + }, + { + "path": "00003367.wav", + "question": "Are there acoustic_guitar and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22997 + }, + { + "path": "00008414.wav", + "question": "Is the trumpet louder than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22998 + }, + { + "path": "00002997.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 22999 + }, + { + "path": "00008145.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23000 + }, + { + "path": "00007225.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23001 + }, + { + "path": "00005451.wav", + "question": "Are there clarinet and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23002 + }, + { + "path": "00005205.wav", + "question": "Are there violin and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23003 + }, + { + "path": "00003931.wav", + "question": "Are there flute and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23004 + }, + { + "path": "00003676.wav", + "question": "Are there tuba and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23005 + }, + { + "path": "00002478.wav", + "question": "Are there piano and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23006 + }, + { + "path": "00004904.wav", + "question": "Are there piano and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23007 + }, + { + "path": "00004746.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23008 + }, + { + "path": "00008218.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23009 + }, + { + "path": "00007809.wav", + "question": "Are there piano and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23010 + }, + { + "path": "00007978.wav", + "question": "Are there saxophone and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23011 + }, + { + "path": "00003127.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23012 + }, + { + "path": "00006948.wav", + "question": "Are there violin and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23013 + }, + { + "path": "00000539.wav", + "question": "Are there flute and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23014 + }, + { + "path": "00002040.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23015 + }, + { + "path": "00008688.wav", + "question": "Are there bassoon and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23016 + }, + { + "path": "00002910.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23017 + }, + { + "path": "00001796.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23018 + }, + { + "path": "00007678.wav", + "question": "Are there flute and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23019 + }, + { + "path": "00008234.wav", + "question": "Are there saxophone and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23020 + }, + { + "path": "00003449.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23021 + }, + { + "path": "00007376.wav", + "question": "Is the cello louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23022 + }, + { + "path": "00005125.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23023 + }, + { + "path": "00000995.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23024 + }, + { + "path": "00005137.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23025 + }, + { + "path": "00007414.wav", + "question": "Is there a guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23026 + }, + { + "path": "00000152.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23027 + }, + { + "path": "00005797.wav", + "question": "Is there a suona sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23028 + }, + { + "path": "00002931.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23029 + }, + { + "path": "00003139.wav", + "question": "Are there piano and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23030 + }, + { + "path": "00008871.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23031 + }, + { + "path": "00008034.wav", + "question": "Are there bassoon and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23032 + }, + { + "path": "00002723.wav", + "question": "Is there a tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23033 + }, + { + "path": "00004797.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23034 + }, + { + "path": "00001762.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23035 + }, + { + "path": "00000002.wav", + "question": "Are there flute and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23036 + }, + { + "path": "00006028.wav", + "question": "Is there a electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23037 + }, + { + "path": "00005860.wav", + "question": "Are there electric_bass and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23038 + }, + { + "path": "00006263.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23039 + }, + { + "path": "00005767.wav", + "question": "Are there piano and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23040 + }, + { + "path": "00000034.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23041 + }, + { + "path": "00006303.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23042 + }, + { + "path": "00005743.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23043 + }, + { + "path": "00003856.wav", + "question": "Are there saxophone and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23044 + }, + { + "path": "00000208.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23045 + }, + { + "path": "00007475.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23046 + }, + { + "path": "00004100.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23047 + }, + { + "path": "00002274.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23048 + }, + { + "path": "00008957.wav", + "question": "Are there clarinet and bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23049 + }, + { + "path": "00005539.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23050 + }, + { + "path": "00001387.wav", + "question": "Are there flute and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23051 + }, + { + "path": "00005298.wav", + "question": "Is the piano playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23052 + }, + { + "path": "00008602.wav", + "question": "Is there a acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23053 + }, + { + "path": "00008548.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23054 + }, + { + "path": "00005926.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23055 + }, + { + "path": "00005153.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23056 + }, + { + "path": "00001099.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23057 + }, + { + "path": "00003033.wav", + "question": "Is the flute playing longer than the suona?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23058 + }, + { + "path": "00002537.wav", + "question": "Is there a clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23059 + }, + { + "path": "00004489.wav", + "question": "Is there a bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23060 + }, + { + "path": "00003630.wav", + "question": "Are there cello and tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23061 + }, + { + "path": "00008750.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23062 + }, + { + "path": "00000907.wav", + "question": "Are there banjo and electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23063 + }, + { + "path": "00004815.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23064 + }, + { + "path": "00001402.wav", + "question": "Is the accordion playing longer than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23065 + }, + { + "path": "00007147.wav", + "question": "Is the piano playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23066 + }, + { + "path": "00003181.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23067 + }, + { + "path": "00003160.wav", + "question": "Are there acoustic_guitar and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23068 + }, + { + "path": "00000486.wav", + "question": "Are there flute and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23069 + }, + { + "path": "00000528.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23070 + }, + { + "path": "00004068.wav", + "question": "Is there a banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23071 + }, + { + "path": "00008645.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23072 + }, + { + "path": "00002349.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23073 + }, + { + "path": "00006413.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23074 + }, + { + "path": "00001052.wav", + "question": "Are there tuba and trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23075 + }, + { + "path": "00002387.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23076 + }, + { + "path": "00006530.wav", + "question": "Are there electric_bass and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23077 + }, + { + "path": "00002765.wav", + "question": "Are there piano and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23078 + }, + { + "path": "00000510.wav", + "question": "Are there violin and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23079 + }, + { + "path": "00008109.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23080 + }, + { + "path": "00005916.wav", + "question": "Are there electric_bass and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23081 + }, + { + "path": "00003829.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23082 + }, + { + "path": "00005559.wav", + "question": "Are there violin and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23083 + }, + { + "path": "00003661.wav", + "question": "Is there a bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23084 + }, + { + "path": "00006134.wav", + "question": "Is there a electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23085 + }, + { + "path": "00006752.wav", + "question": "Is there a electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23086 + }, + { + "path": "00005075.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23087 + }, + { + "path": "00006513.wav", + "question": "Is there a banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23088 + }, + { + "path": "00002160.wav", + "question": "Is there a guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23089 + }, + { + "path": "00001081.wav", + "question": "Are there saxophone and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23090 + }, + { + "path": "00006047.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23091 + }, + { + "path": "00006091.wav", + "question": "Are there tuba and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23092 + }, + { + "path": "00006583.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23093 + }, + { + "path": "00002560.wav", + "question": "Are there violin and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23094 + }, + { + "path": "00004647.wav", + "question": "Are there electric_bass and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23095 + }, + { + "path": "00004422.wav", + "question": "Are there flute and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23096 + }, + { + "path": "00008054.wav", + "question": "Are there trumpet and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23097 + }, + { + "path": "00008315.wav", + "question": "Are there trumpet and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23098 + }, + { + "path": "00003431.wav", + "question": "Are there acoustic_guitar and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23099 + }, + { + "path": "00008884.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23100 + }, + { + "path": "00002678.wav", + "question": "Is there a bagpipe sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23101 + }, + { + "path": "00002252.wav", + "question": "Is the cello louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23102 + }, + { + "path": "00005312.wav", + "question": "Are there ukulele and bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23103 + }, + { + "path": "00004982.wav", + "question": "Is there a pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23104 + }, + { + "path": "00007097.wav", + "question": "Are there bagpipe and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23105 + }, + { + "path": "00006392.wav", + "question": "Are there violin and erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23106 + }, + { + "path": "00006809.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23107 + }, + { + "path": "00006795.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23108 + }, + { + "path": "00001844.wav", + "question": "Are there trumpet and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23109 + }, + { + "path": "00000696.wav", + "question": "Are there electric_bass and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23110 + }, + { + "path": "00006373.wav", + "question": "Are there acoustic_guitar and tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23111 + }, + { + "path": "00007941.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23112 + }, + { + "path": "00001626.wav", + "question": "Are there violin and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23113 + }, + { + "path": "00000172.wav", + "question": "Is the cello playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23114 + }, + { + "path": "00007797.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23115 + }, + { + "path": "00005885.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23116 + }, + { + "path": "00005271.wav", + "question": "Are there banjo and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23117 + }, + { + "path": "00004779.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23118 + }, + { + "path": "00000069.wav", + "question": "Are there accordion and guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23119 + }, + { + "path": "00004214.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23120 + }, + { + "path": "00008489.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23121 + }, + { + "path": "00001702.wav", + "question": "Are there drum and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23122 + }, + { + "path": "00000975.wav", + "question": "Are there piano and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23123 + }, + { + "path": "00004866.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23124 + }, + { + "path": "00004667.wav", + "question": "Is there a erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23125 + }, + { + "path": "00008129.wav", + "question": "Is there a piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23126 + }, + { + "path": "00005115.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23127 + }, + { + "path": "00008474.wav", + "question": "Is there a violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23128 + }, + { + "path": "00002578.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23129 + }, + { + "path": "00007268.wav", + "question": "Are there acoustic_guitar and ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23130 + }, + { + "path": "00006147.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23131 + }, + { + "path": "00002872.wav", + "question": "Is the piano more rhythmic than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23132 + }, + { + "path": "00004318.wav", + "question": "Are there ukulele and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23133 + }, + { + "path": "00007318.wav", + "question": "Are there saxophone and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23134 + }, + { + "path": "00001540.wav", + "question": "Are there ukulele and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23135 + }, + { + "path": "00006343.wav", + "question": "Is there a saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23136 + }, + { + "path": "00008517.wav", + "question": "Are there erhu and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23137 + }, + { + "path": "00002176.wav", + "question": "Is the saxophone louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23138 + }, + { + "path": "00006385.wav", + "question": "Are there violin and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23139 + }, + { + "path": "00002312.wav", + "question": "Are there flute and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23140 + }, + { + "path": "00008896.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23141 + }, + { + "path": "00001522.wav", + "question": "Are there clarinet and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23142 + }, + { + "path": "00007066.wav", + "question": "Is the cello playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23143 + }, + { + "path": "00008845.wav", + "question": "Are there acoustic_guitar and bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23144 + }, + { + "path": "00001608.wav", + "question": "Are there saxophone and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23145 + }, + { + "path": "00004837.wav", + "question": "Are there flute and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23146 + }, + { + "path": "00008720.wav", + "question": "Is there a xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23147 + }, + { + "path": "00006773.wav", + "question": "Are there erhu and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23148 + }, + { + "path": "00008202.wav", + "question": "Are there erhu and guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23149 + }, + { + "path": "00006724.wav", + "question": "Are there banjo and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23150 + }, + { + "path": "00006285.wav", + "question": "Is there a trumpet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23151 + }, + { + "path": "00001147.wav", + "question": "Is there a bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23152 + }, + { + "path": "00001125.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23153 + }, + { + "path": "00005286.wav", + "question": "Are there piano and erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23154 + }, + { + "path": "00000415.wav", + "question": "Are there flute and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23155 + }, + { + "path": "00001186.wav", + "question": "Is there a flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23156 + }, + { + "path": "00002259.wav", + "question": "Are there violin and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23157 + }, + { + "path": "00005899.wav", + "question": "Is there a accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23158 + }, + { + "path": "00003754.wav", + "question": "Are there saxophone and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23159 + }, + { + "path": "00006491.wav", + "question": "Is there a ukulele sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23160 + }, + { + "path": "00007438.wav", + "question": "Is there a drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23161 + }, + { + "path": "00001365.wav", + "question": "Is there a cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23162 + }, + { + "path": "00004711.wav", + "question": "Are there accordion and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23163 + }, + { + "path": "00004944.wav", + "question": "Are there piano and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23164 + }, + { + "path": "00000573.wav", + "question": "Are there clarinet and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23165 + }, + { + "path": "00007247.wav", + "question": "Are there piano and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23166 + }, + { + "path": "00006629.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23167 + }, + { + "path": "00003598.wav", + "question": "Are there acoustic_guitar and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23168 + }, + { + "path": "00003316.wav", + "question": "Are there tuba and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23169 + }, + { + "path": "00005497.wav", + "question": "Is the trumpet louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23170 + }, + { + "path": "00007698.wav", + "question": "Are there pipa and erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23171 + }, + { + "path": "00006574.wav", + "question": "Are there violin and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23172 + }, + { + "path": "00007018.wav", + "question": "Are there violin and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23173 + }, + { + "path": "00002771.wav", + "question": "Is the accordion louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23174 + }, + { + "path": "00001644.wav", + "question": "Are there saxophone and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23175 + }, + { + "path": "00002335.wav", + "question": "Are there violin and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23176 + }, + { + "path": "00001776.wav", + "question": "Are there clarinet and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23177 + }, + { + "path": "00003615.wav", + "question": "Are there acoustic_guitar and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23178 + }, + { + "path": "00006659.wav", + "question": "Are there accordion and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23179 + }, + { + "path": "00006995.wav", + "question": "Are there violin and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23180 + }, + { + "path": "00000287.wav", + "question": "Are there clarinet and bassoon sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23181 + }, + { + "path": "00000643.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23182 + }, + { + "path": "00004686.wav", + "question": "Is the flute playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23183 + }, + { + "path": "00003193.wav", + "question": "Are there piano and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23184 + }, + { + "path": "00005057.wav", + "question": "Are there flute and accordion sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23185 + }, + { + "path": "00007296.wav", + "question": "Are there acoustic_guitar and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23186 + }, + { + "path": "00004343.wav", + "question": "Are there violin and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23187 + }, + { + "path": "00008514.wav", + "question": "Is the flute louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23188 + }, + { + "path": "00000312.wav", + "question": "Are there drum and congas sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23189 + }, + { + "path": "00000048.wav", + "question": "Are there guzheng and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23190 + }, + { + "path": "00007073.wav", + "question": "Are there trumpet and cello sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23191 + }, + { + "path": "00003777.wav", + "question": "Are there clarinet and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23192 + }, + { + "path": "00000940.wav", + "question": "Are there acoustic_guitar and banjo sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23193 + }, + { + "path": "00000685.wav", + "question": "Are there drum and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23194 + }, + { + "path": "00007514.wav", + "question": "Are there cello and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23195 + }, + { + "path": "00008359.wav", + "question": "Are there acoustic_guitar and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23196 + }, + { + "path": "00008845.wav", + "question": "Is the bassoon playing longer than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23197 + }, + { + "path": "00001291.wav", + "question": "Are there accordion and tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23198 + }, + { + "path": "00007717.wav", + "question": "Are there ukulele and flute sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23199 + }, + { + "path": "00006479.wav", + "question": "Are there pipa and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23200 + }, + { + "path": "00002801.wav", + "question": "Are there piano and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23201 + }, + { + "path": "00004385.wav", + "question": "Are there piano and saxophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23202 + }, + { + "path": "00007832.wav", + "question": "Are there pipa and congas sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23203 + }, + { + "path": "00002179.wav", + "question": "Are there cello and piano sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23204 + }, + { + "path": "00000746.wav", + "question": "Are there electric_bass and congas sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23205 + }, + { + "path": "00005521.wav", + "question": "Are there flute and guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23206 + }, + { + "path": "00003520.wav", + "question": "Are there acoustic_guitar and pipa sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23207 + }, + { + "path": "00001800.wav", + "question": "Is the piano more rhythmic than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23208 + }, + { + "path": "00008586.wav", + "question": "Is the flute playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23209 + }, + { + "path": "00004029.wav", + "question": "Are there piano and tuba sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23210 + }, + { + "path": "00004537.wav", + "question": "Is the acoustic_guitar louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23211 + }, + { + "path": "00004385.wav", + "question": "Is the piano louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23212 + }, + { + "path": "00003101.wav", + "question": "Are there bassoon and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23213 + }, + { + "path": "00004133.wav", + "question": "Are there erhu and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23214 + }, + { + "path": "00008579.wav", + "question": "Are there saxophone and electric_bass sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23215 + }, + { + "path": "00004275.wav", + "question": "Are there saxophone and drum sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23216 + }, + { + "path": "00001320.wav", + "question": "Are there saxophone and clarinet sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23217 + }, + { + "path": "00006216.wav", + "question": "Are there banjo and guzheng sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23218 + }, + { + "path": "00000350.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23219 + }, + { + "path": "00005915.wav", + "question": "Is the cello more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23220 + }, + { + "path": "00001337.wav", + "question": "Is the clarinet playing longer than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23221 + }, + { + "path": "00002346.wav", + "question": "Is the violin more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23222 + }, + { + "path": "00003887.wav", + "question": "Is the accordion playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23223 + }, + { + "path": "00007375.wav", + "question": "Are there flute and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23224 + }, + { + "path": "00001905.wav", + "question": "Are there saxophone and acoustic_guitar sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23225 + }, + { + "path": "00001147.wav", + "question": "Is the bassoon playing longer than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23226 + }, + { + "path": "00001678.wav", + "question": "Is the acoustic_guitar playing longer than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23227 + }, + { + "path": "00006365.wav", + "question": "Is the erhu louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23228 + }, + { + "path": "00002057.wav", + "question": "Are there cello and suona sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23229 + }, + { + "path": "00007353.wav", + "question": "Are there guzheng and erhu sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23230 + }, + { + "path": "00001412.wav", + "question": "Are there pipa and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23231 + }, + { + "path": "00007488.wav", + "question": "Are there trumpet and xylophone sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23232 + }, + { + "path": "00008671.wav", + "question": "Are there clarinet and violin sound?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23233 + }, + { + "path": "00007770.wav", + "question": "Is the acoustic_guitar more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23234 + }, + { + "path": "00004261.wav", + "question": "Is the cello more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23235 + }, + { + "path": "00005138.wav", + "question": "Is the cello louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23236 + }, + { + "path": "00003503.wav", + "question": "Is the violin louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23237 + }, + { + "path": "00008742.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23238 + }, + { + "path": "sa00008119.wav", + "question": "Is the bagpipe playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23239 + }, + { + "path": "00006118.wav", + "question": "Is the acoustic_guitar louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23240 + }, + { + "path": "00007994.wav", + "question": "Is the cello playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23241 + }, + { + "path": "00006501.wav", + "question": "Is the saxophone more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23242 + }, + { + "path": "00004448.wav", + "question": "Is the clarinet louder than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23243 + }, + { + "path": "00000659.wav", + "question": "Is the acoustic_guitar louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23244 + }, + { + "path": "00002051.wav", + "question": "Is the cello playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23245 + }, + { + "path": "sa00007433.wav", + "question": "Is the guzheng louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23246 + }, + { + "path": "00005997.wav", + "question": "Is the saxophone playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23247 + }, + { + "path": "00004939.wav", + "question": "Is the violin louder than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23248 + }, + { + "path": "va00005964.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23249 + }, + { + "path": "va00005379.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23250 + }, + { + "path": "va00005596.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23251 + }, + { + "path": "00002628.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "four", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23252 + }, + { + "path": "sa00007149.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23253 + }, + { + "path": "00007555.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23254 + }, + { + "path": "00001240.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23255 + }, + { + "path": "00002294.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23256 + }, + { + "path": "00000467.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "four", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23257 + }, + { + "path": "sa00007562.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23258 + }, + { + "path": "vv00008937.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23259 + }, + { + "path": "00008974.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23260 + }, + { + "path": "00001862.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23261 + }, + { + "path": "00001882.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23262 + }, + { + "path": "00004759.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23263 + }, + { + "path": "va00004777.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23264 + }, + { + "path": "va00004777.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23265 + }, + { + "path": "00002781.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23266 + }, + { + "path": "00008623.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23267 + }, + { + "path": "00003290.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "three", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23268 + }, + { + "path": "va00003539.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23269 + }, + { + "path": "va00005462.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23270 + }, + { + "path": "00005169.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23271 + }, + { + "path": "vv00000088.wav", + "question": "Is the tuba more rhythmic than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23272 + }, + { + "path": "va00005111.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "three", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23273 + }, + { + "path": "00000952.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "four", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23274 + }, + { + "path": "00001669.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "four", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23275 + }, + { + "path": "00004082.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23276 + }, + { + "path": "vv00000531.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23277 + }, + { + "path": "sa00007332.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23278 + }, + { + "path": "00006070.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23279 + }, + { + "path": "sa00008398.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23280 + }, + { + "path": "00003907.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23281 + }, + { + "path": "00007639.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23282 + }, + { + "path": "va00003153.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23283 + }, + { + "path": "00002822.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23284 + }, + { + "path": "00000926.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23285 + }, + { + "path": "00002895.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23286 + }, + { + "path": "00002657.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23287 + }, + { + "path": "00008535.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23288 + }, + { + "path": "00003481.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23289 + }, + { + "path": "vv00000268.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23290 + }, + { + "path": "00002447.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23291 + }, + { + "path": "00007470.wav", + "question": "Is the bassoon playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23292 + }, + { + "path": "00001309.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23293 + }, + { + "path": "00000825.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23294 + }, + { + "path": "00008411.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "five", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23295 + }, + { + "path": "00004197.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23296 + }, + { + "path": "00003235.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23297 + }, + { + "path": "00005684.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23298 + }, + { + "path": "00000785.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "more than ten", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "more than ten", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23299 + }, + { + "path": "00006168.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23300 + }, + { + "path": "va00003836.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23301 + }, + { + "path": "00004442.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23302 + }, + { + "path": "00004570.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23303 + }, + { + "path": "va00005770.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23304 + }, + { + "path": "00008452.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23305 + }, + { + "path": "00007164.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23306 + }, + { + "path": "00008561.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23307 + }, + { + "path": "00007959.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23308 + }, + { + "path": "00004371.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23309 + }, + { + "path": "00002021.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23310 + }, + { + "path": "00000350.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23311 + }, + { + "path": "00002516.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23312 + }, + { + "path": "va00004905.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23313 + }, + { + "path": "00008913.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23314 + }, + { + "path": "00001265.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23315 + }, + { + "path": "sa00006762.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23316 + }, + { + "path": "sa00006534.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23317 + }, + { + "path": "sa00008030.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23318 + }, + { + "path": "00002424.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23319 + }, + { + "path": "00006230.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23320 + }, + { + "path": "va00004025.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23321 + }, + { + "path": "vv00002616.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23322 + }, + { + "path": "00006115.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23323 + }, + { + "path": "00007572.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23324 + }, + { + "path": "00000371.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23325 + }, + { + "path": "00004922.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23326 + }, + { + "path": "00003559.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23327 + }, + { + "path": "00008082.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23328 + }, + { + "path": "00008204.wav", + "question": "Is the ukulele more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23329 + }, + { + "path": "00002198.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23330 + }, + { + "path": "sa00008869.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23331 + }, + { + "path": "00004004.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23332 + }, + { + "path": "00001929.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23333 + }, + { + "path": "00006321.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23334 + }, + { + "path": "00005818.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "six", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "six", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23335 + }, + { + "path": "00006461.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23336 + }, + { + "path": "00003332.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "three", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23337 + }, + { + "path": "va00005299.wav", + "question": "Is the clarinet louder than the congas?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23338 + }, + { + "path": "00004233.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "three", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23339 + }, + { + "path": "sa00008741.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23340 + }, + { + "path": "00004298.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23341 + }, + { + "path": "00005031.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23342 + }, + { + "path": "00007917.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23343 + }, + { + "path": "00007994.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23344 + }, + { + "path": "00003541.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23345 + }, + { + "path": "00006908.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23346 + }, + { + "path": "00002235.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "three", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23347 + }, + { + "path": "00001160.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23348 + }, + { + "path": "00002363.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23349 + }, + { + "path": "00002363.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "five", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "five", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23350 + }, + { + "path": "00000888.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23351 + }, + { + "path": "00005839.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23352 + }, + { + "path": "va00005320.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23353 + }, + { + "path": "00008708.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23354 + }, + { + "path": "00007843.wav", + "question": "Is the piano playing longer than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23355 + }, + { + "path": "00007616.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "four", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23356 + }, + { + "path": "sa00007435.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23357 + }, + { + "path": "00006970.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23358 + }, + { + "path": "00005399.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23359 + }, + { + "path": "00004533.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23360 + }, + { + "path": "00007124.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23361 + }, + { + "path": "va00004284.wav", + "question": "Is the erhu louder than the xylophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23362 + }, + { + "path": "00001985.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23363 + }, + { + "path": "00003737.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23364 + }, + { + "path": "sa00006260.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23365 + }, + { + "path": "va00004294.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23366 + }, + { + "path": "sa00006859.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23367 + }, + { + "path": "00004855.wav", + "question": "Is the cello louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23368 + }, + { + "path": "va00003356.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23369 + }, + { + "path": "vv00000177.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23370 + }, + { + "path": "va00003268.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23371 + }, + { + "path": "00008821.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23372 + }, + { + "path": "00005436.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23373 + }, + { + "path": "sa00006195.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23374 + }, + { + "path": "va00003919.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23375 + }, + { + "path": "00002492.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23376 + }, + { + "path": "00000846.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23377 + }, + { + "path": "00000242.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23378 + }, + { + "path": "sa00006645.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23379 + }, + { + "path": "00002435.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "three", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23380 + }, + { + "path": "sa00007890.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23381 + }, + { + "path": "va00005220.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23382 + }, + { + "path": "00005632.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "three", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23383 + }, + { + "path": "sa00008136.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23384 + }, + { + "path": "sa00008136.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23385 + }, + { + "path": "00003045.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23386 + }, + { + "path": "00005662.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23387 + }, + { + "path": "00004215.wav", + "question": "Is the saxophone louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23388 + }, + { + "path": "va00005427.wav", + "question": "Is the saxophone louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23389 + }, + { + "path": "00002212.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23390 + }, + { + "path": "00001912.wav", + "question": "Is the pipa louder than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23391 + }, + { + "path": "00002074.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23392 + }, + { + "path": "00006698.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23393 + }, + { + "path": "00003695.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23394 + }, + { + "path": "00003976.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "three", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23395 + }, + { + "path": "00007620.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23396 + }, + { + "path": "00002291.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23397 + }, + { + "path": "va00005024.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23398 + }, + { + "path": "va00005203.wav", + "question": "Is the bagpipe louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23399 + }, + { + "path": "00002705.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23400 + }, + { + "path": "00000039.wav", + "question": "Is the acoustic_guitar louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23401 + }, + { + "path": "00004147.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23402 + }, + { + "path": "sa00008256.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23403 + }, + { + "path": "sa00008646.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23404 + }, + { + "path": "00005371.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23405 + }, + { + "path": "00004128.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23406 + }, + { + "path": "00005362.wav", + "question": "Is the violin more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23407 + }, + { + "path": "00005981.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23408 + }, + { + "path": "vv00000444.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23409 + }, + { + "path": "sa00008524.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23410 + }, + { + "path": "sa00007768.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23411 + }, + { + "path": "00003952.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23412 + }, + { + "path": "00006164.wav", + "question": "Is the acoustic_guitar louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23413 + }, + { + "path": "sa00007027.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "one", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23414 + }, + { + "path": "va00004370.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "four", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "four", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23415 + }, + { + "path": "va00004512.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23416 + }, + { + "path": "va00005862.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23417 + }, + { + "path": "00000624.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23418 + }, + { + "path": "00006746.wav", + "question": "Is the flute louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23419 + }, + { + "path": "vv00000572.wav", + "question": "Is the ukulele louder than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23420 + }, + { + "path": "00003010.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23421 + }, + { + "path": "00004053.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23422 + }, + { + "path": "00001441.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23423 + }, + { + "path": "sa00006863.wav", + "question": "Is the flute louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23424 + }, + { + "path": "00002103.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23425 + }, + { + "path": "00007322.wav", + "question": "Is the saxophone louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23426 + }, + { + "path": "00007322.wav", + "question": "Is the saxophone louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23427 + }, + { + "path": "00001511.wav", + "question": "Is the accordion louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23428 + }, + { + "path": "00005415.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23429 + }, + { + "path": "00000243.wav", + "question": "Is the trumpet louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23430 + }, + { + "path": "00005580.wav", + "question": "Is the cello louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23431 + }, + { + "path": "00002694.wav", + "question": "Is the piano louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23432 + }, + { + "path": "00008252.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "five", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23433 + }, + { + "path": "00002407.wav", + "question": "Is the clarinet louder than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23434 + }, + { + "path": "00003718.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23435 + }, + { + "path": "va00004677.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23436 + }, + { + "path": "00002149.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23437 + }, + { + "path": "00001031.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23438 + }, + { + "path": "va00004157.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23439 + }, + { + "path": "00006419.wav", + "question": "Is the piano louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23440 + }, + { + "path": "sa00007659.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23441 + }, + { + "path": "00004399.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23442 + }, + { + "path": "vv00002421.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23443 + }, + { + "path": "sa00006077.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23444 + }, + { + "path": "00008793.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23445 + }, + { + "path": "00000448.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23446 + }, + { + "path": "00004115.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23447 + }, + { + "path": "va00003446.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23448 + }, + { + "path": "va00003692.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23449 + }, + { + "path": "sa00006937.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23450 + }, + { + "path": "sa00005995.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23451 + }, + { + "path": "sa00007196.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "one", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23452 + }, + { + "path": "00000605.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23453 + }, + { + "path": "va00004580.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23454 + }, + { + "path": "00006688.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23455 + }, + { + "path": "00005415.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23456 + }, + { + "path": "00007109.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "five", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23457 + }, + { + "path": "00002855.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23458 + }, + { + "path": "00000178.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23459 + }, + { + "path": "00005235.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23460 + }, + { + "path": "00004619.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23461 + }, + { + "path": "sa00006454.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23462 + }, + { + "path": "00006441.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23463 + }, + { + "path": "vv00002750.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "one", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23464 + }, + { + "path": "00004967.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23465 + }, + { + "path": "vv00002969.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "two", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23466 + }, + { + "path": "00003413.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "three", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23467 + }, + { + "path": "va00005533.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23468 + }, + { + "path": "00001742.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "two", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "two", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23469 + }, + { + "path": "00006713.wav", + "question": "How many musical instruments were heard throughout the video?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "seven", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23470 + }, + { + "path": "vv00002643.wav", + "question": "Is the trumpet louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23471 + }, + { + "path": "00001242.wav", + "question": "Is the cello louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23472 + }, + { + "path": "va00003244.wav", + "question": "Is the saxophone louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23473 + }, + { + "path": "va00005518.wav", + "question": "Is the saxophone louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23474 + }, + { + "path": "va00004059.wav", + "question": "Is the congas more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23475 + }, + { + "path": "00005810.wav", + "question": "Is the acoustic_guitar louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23476 + }, + { + "path": "sa00006984.wav", + "question": "Is the saxophone louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23477 + }, + { + "path": "vv00000190.wav", + "question": "Is the piano louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23478 + }, + { + "path": "va00003982.wav", + "question": "Is the suona playing longer than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23479 + }, + { + "path": "00004788.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23480 + }, + { + "path": "00006689.wav", + "question": "Is the piano louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23481 + }, + { + "path": "00007219.wav", + "question": "Is the acoustic_guitar more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23482 + }, + { + "path": "00004347.wav", + "question": "Is the piano louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23483 + }, + { + "path": "00001575.wav", + "question": "Is the piano louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23484 + }, + { + "path": "va00003533.wav", + "question": "Is the saxophone playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23485 + }, + { + "path": "00002522.wav", + "question": "Is the erhu louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23486 + }, + { + "path": "vv00002914.wav", + "question": "Is the guzheng louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23487 + }, + { + "path": "vv00000323.wav", + "question": "Is the guzheng louder than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23488 + }, + { + "path": "va00005060.wav", + "question": "Is the xylophone louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23489 + }, + { + "path": "va00003456.wav", + "question": "Is the accordion louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23490 + }, + { + "path": "va00003675.wav", + "question": "Is the drum louder than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23491 + }, + { + "path": "vv00003027.wav", + "question": "Is the bagpipe louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23492 + }, + { + "path": "va00005827.wav", + "question": "Is the xylophone louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23493 + }, + { + "path": "va00003376.wav", + "question": "Is the congas louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23494 + }, + { + "path": "00001031.wav", + "question": "Is the erhu louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23495 + }, + { + "path": "sa00006083.wav", + "question": "Is the saxophone louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23496 + }, + { + "path": "sa00005991.wav", + "question": "Is the pipa louder than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23497 + }, + { + "path": "sa00006210.wav", + "question": "Is the saxophone louder than the congas?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23498 + }, + { + "path": "va00005723.wav", + "question": "Is the saxophone louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23499 + }, + { + "path": "sa00008817.wav", + "question": "Is the erhu louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23500 + }, + { + "path": "00007948.wav", + "question": "Is the violin louder than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23501 + }, + { + "path": "00000314.wav", + "question": "Is the congas louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23502 + }, + { + "path": "00001628.wav", + "question": "Is the violin louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23503 + }, + { + "path": "00001057.wav", + "question": "Is the trumpet more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23504 + }, + { + "path": "sa00008229.wav", + "question": "Is the drum louder than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23505 + }, + { + "path": "00000812.wav", + "question": "Is the piano more rhythmic than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23506 + }, + { + "path": "sa00007103.wav", + "question": "Is the suona louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23507 + }, + { + "path": "00000876.wav", + "question": "Is the saxophone louder than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23508 + }, + { + "path": "sa00006341.wav", + "question": "Is the xylophone playing longer than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23509 + }, + { + "path": "va00004513.wav", + "question": "Is the banjo louder than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23510 + }, + { + "path": "va00004591.wav", + "question": "Is the drum louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23511 + }, + { + "path": "va00005894.wav", + "question": "Is the piano louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23512 + }, + { + "path": "sa00007323.wav", + "question": "Is the xylophone louder than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23513 + }, + { + "path": "00008939.wav", + "question": "Is the accordion louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23514 + }, + { + "path": "00003765.wav", + "question": "Is the drum louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23515 + }, + { + "path": "va00003853.wav", + "question": "Is the drum more rhythmic than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23516 + }, + { + "path": "sa00008513.wav", + "question": "Is the trumpet playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23517 + }, + { + "path": "sa00008314.wav", + "question": "Is the bagpipe more rhythmic than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23518 + }, + { + "path": "sa00007799.wav", + "question": "Is the drum playing longer than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23519 + }, + { + "path": "va00003115.wav", + "question": "Is the guzheng more rhythmic than the bagpipe?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23520 + }, + { + "path": "sa00006712.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23521 + }, + { + "path": "eva00000772.wav", + "question": "Is the saxophone playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23522 + }, + { + "path": "eva00001021.wav", + "question": "Is the accordion louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23523 + }, + { + "path": "evv00001324.wav", + "question": "Is the piano more rhythmic than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23524 + }, + { + "path": "eva00000818.wav", + "question": "Is the clarinet more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23525 + }, + { + "path": "esa00000652.wav", + "question": "Is the erhu louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23526 + }, + { + "path": "eva00000718.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23527 + }, + { + "path": "esa00000615.wav", + "question": "Is the trumpet louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23528 + }, + { + "path": "esa00000540.wav", + "question": "Is the piano more rhythmic than the suona?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23529 + }, + { + "path": "evv00001228.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23530 + }, + { + "path": "eva00000951.wav", + "question": "Is the clarinet louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23531 + }, + { + "path": "evv00001407.wav", + "question": "Is the cello louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23532 + }, + { + "path": "esa00000291.wav", + "question": "Is the acoustic_guitar louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23533 + }, + { + "path": "eva00001089.wav", + "question": "Is the flute louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23534 + }, + { + "path": "eva00000875.wav", + "question": "Is the violin more rhythmic than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23535 + }, + { + "path": "eva00000875.wav", + "question": "Is the violin more rhythmic than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23536 + }, + { + "path": "eva00000980.wav", + "question": "Is the piano playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23537 + }, + { + "path": "eva00001005.wav", + "question": "Is the guzheng louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23538 + }, + { + "path": "eva00001189.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23539 + }, + { + "path": "esa00000660.wav", + "question": "Is the drum more rhythmic than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23540 + }, + { + "path": "eva00000783.wav", + "question": "Is the tuba louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23541 + }, + { + "path": "eva00001201.wav", + "question": "Is the flute louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23542 + }, + { + "path": "00008047.wav", + "question": "Is the acoustic_guitar more rhythmic than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23543 + }, + { + "path": "00005069.wav", + "question": "Is the violin louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23544 + }, + { + "path": "00003699.wav", + "question": "Is the cello more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23545 + }, + { + "path": "evv00001397.wav", + "question": "Is the piano playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23546 + }, + { + "path": "esa00000259.wav", + "question": "Is the acoustic_guitar playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23547 + }, + { + "path": "evv00001243.wav", + "question": "Is the clarinet louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23548 + }, + { + "path": "esa00000522.wav", + "question": "Is the bassoon louder than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23549 + }, + { + "path": "esa00000397.wav", + "question": "Is the violin louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23550 + }, + { + "path": "evv00001381.wav", + "question": "Is the violin louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23551 + }, + { + "path": "evv00001278.wav", + "question": "Is the acoustic_guitar more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23552 + }, + { + "path": "eva00001032.wav", + "question": "Is the violin louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23553 + }, + { + "path": "eva00001182.wav", + "question": "Is the piano playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23554 + }, + { + "path": "vv00002764.wav", + "question": "Is the saxophone louder than the suona?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23555 + }, + { + "path": "eva00000970.wav", + "question": "Is the clarinet louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23556 + }, + { + "path": "esa00000426.wav", + "question": "Is the banjo more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23557 + }, + { + "path": "evv00001387.wav", + "question": "Is the flute playing longer than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23558 + }, + { + "path": "eva00000975.wav", + "question": "Is the piano louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23559 + }, + { + "path": "00001203.wav", + "question": "Is the flute louder than the banjo?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23560 + }, + { + "path": "esa00000620.wav", + "question": "Is the piano louder than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23561 + }, + { + "path": "eva00000732.wav", + "question": "Is the saxophone louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23562 + }, + { + "path": "evv00001481.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23563 + }, + { + "path": "eva00000756.wav", + "question": "Is the violin more rhythmic than the banjo?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23564 + }, + { + "path": "evv00001252.wav", + "question": "Is the piano playing longer than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23565 + }, + { + "path": "00006892.wav", + "question": "Is the flute louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23566 + }, + { + "path": "00006892.wav", + "question": "Is the flute louder than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23567 + }, + { + "path": "eva00001099.wav", + "question": "Is the drum more rhythmic than the electric_bass?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23568 + }, + { + "path": "eva00000726.wav", + "question": "Is the acoustic_guitar more rhythmic than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23569 + }, + { + "path": "esa00000149.wav", + "question": "Is the flute playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23570 + }, + { + "path": "00004899.wav", + "question": "Is the violin louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23571 + }, + { + "path": "esa00000493.wav", + "question": "Is the acoustic_guitar louder than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23572 + }, + { + "path": "eva00000805.wav", + "question": "Is the violin more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23573 + }, + { + "path": "esa00000549.wav", + "question": "Is the violin louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23574 + }, + { + "path": "eva00001118.wav", + "question": "Is the bassoon more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23575 + }, + { + "path": "00003103.wav", + "question": "Is the electric_bass louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23576 + }, + { + "path": "esa00000277.wav", + "question": "Is the acoustic_guitar more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23577 + }, + { + "path": "eva00000962.wav", + "question": "Is the violin playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23578 + }, + { + "path": "eva00000832.wav", + "question": "Is the saxophone playing longer than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23579 + }, + { + "path": "esa00000559.wav", + "question": "Is the cello louder than the congas?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23580 + }, + { + "path": "evv00001366.wav", + "question": "Is the saxophone louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23581 + }, + { + "path": "evv00001353.wav", + "question": "Is the acoustic_guitar playing longer than the banjo?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23582 + }, + { + "path": "eva00000811.wav", + "question": "Is the tuba playing longer than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23583 + }, + { + "path": "00006314.wav", + "question": "Is the bagpipe louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23584 + }, + { + "path": "eva00000697.wav", + "question": "Is the acoustic_guitar more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23585 + }, + { + "path": "esa00000301.wav", + "question": "Is the violin more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23586 + }, + { + "path": "eva00000932.wav", + "question": "Is the piano playing longer than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23587 + }, + { + "path": "eva00000903.wav", + "question": "Is the saxophone louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23588 + }, + { + "path": "00008328.wav", + "question": "Is the cello more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23589 + }, + { + "path": "eva00000825.wav", + "question": "Is the pipa more rhythmic than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23590 + }, + { + "path": "esa00000296.wav", + "question": "Is the piano louder than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23591 + }, + { + "path": "00008683.wav", + "question": "Is the piano louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23592 + }, + { + "path": "eva00001132.wav", + "question": "Is the clarinet louder than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23593 + }, + { + "path": "esa00000475.wav", + "question": "Is the clarinet more rhythmic than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23594 + }, + { + "path": "eva00000705.wav", + "question": "Is the banjo more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23595 + }, + { + "path": "eva00000705.wav", + "question": "Is the banjo more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23596 + }, + { + "path": "esa00000199.wav", + "question": "Is the accordion more rhythmic than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23597 + }, + { + "path": "00003286.wav", + "question": "Is the violin playing longer than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23598 + }, + { + "path": "evv00001304.wav", + "question": "Is the tuba louder than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23599 + }, + { + "path": "esa00000565.wav", + "question": "Is the accordion louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23600 + }, + { + "path": "esa00000601.wav", + "question": "Is the piano louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23601 + }, + { + "path": "esa00000628.wav", + "question": "Is the acoustic_guitar more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23602 + }, + { + "path": "esa00000313.wav", + "question": "Is the flute more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23603 + }, + { + "path": "eva00001215.wav", + "question": "Is the flute playing longer than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23604 + }, + { + "path": "eva00000911.wav", + "question": "Is the acoustic_guitar playing longer than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23605 + }, + { + "path": "esa00000232.wav", + "question": "Is the acoustic_guitar more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23606 + }, + { + "path": "evv00001442.wav", + "question": "Is the xylophone more rhythmic than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23607 + }, + { + "path": "eva00001205.wav", + "question": "Is the accordion louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23608 + }, + { + "path": "eva00001166.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23609 + }, + { + "path": "esa00000328.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23610 + }, + { + "path": "eva00001145.wav", + "question": "Is the acoustic_guitar playing longer than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23611 + }, + { + "path": "eva00000987.wav", + "question": "Is the violin louder than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23612 + }, + { + "path": "esa00000402.wav", + "question": "Is the acoustic_guitar playing longer than the banjo?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23613 + }, + { + "path": "eva00001081.wav", + "question": "Is the piano louder than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23614 + }, + { + "path": "eva00000891.wav", + "question": "Is the clarinet louder than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23615 + }, + { + "path": "eva00001173.wav", + "question": "Is the electric_bass playing longer than the accordion?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23616 + }, + { + "path": "eva00001063.wav", + "question": "Is the banjo playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23617 + }, + { + "path": "eva00001070.wav", + "question": "Is the erhu more rhythmic than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23618 + }, + { + "path": "eva00000921.wav", + "question": "Is the banjo playing longer than the suona?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23619 + }, + { + "path": "00008080.wav", + "question": "Is the xylophone louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23620 + }, + { + "path": "sa00007710.wav", + "question": "Is the acoustic_guitar playing longer than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23621 + }, + { + "path": "evv00001292.wav", + "question": "Is the violin more rhythmic than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23622 + }, + { + "path": "esa00000442.wav", + "question": "Is the acoustic_guitar more rhythmic than the ukulele?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23623 + }, + { + "path": "evv00001334.wav", + "question": "Is the ukulele more rhythmic than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23624 + }, + { + "path": "evv00001454.wav", + "question": "Is the violin more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23625 + }, + { + "path": "eva00001156.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23626 + }, + { + "path": "eva00000866.wav", + "question": "Is the electric_bass louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23627 + }, + { + "path": "evv00001343.wav", + "question": "Is the suona louder than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23628 + }, + { + "path": "evv00001427.wav", + "question": "Is the clarinet louder than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23629 + }, + { + "path": "eva00001114.wav", + "question": "Is the piano playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23630 + }, + { + "path": "va00004989.wav", + "question": "Is the xylophone louder than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23631 + }, + { + "path": "00006266.wav", + "question": "Is the violin playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23632 + }, + { + "path": "eva00000792.wav", + "question": "Is the clarinet playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23633 + }, + { + "path": "evv00001264.wav", + "question": "Is the guzheng more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23634 + }, + { + "path": "00002615.wav", + "question": "Is the pipa more rhythmic than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23635 + }, + { + "path": "evv00001415.wav", + "question": "Is the erhu more rhythmic than the suona?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23636 + }, + { + "path": "eva00000748.wav", + "question": "Is the trumpet more rhythmic than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23637 + }, + { + "path": "eva00001044.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23638 + }, + { + "path": "00006636.wav", + "question": "Is the accordion more rhythmic than the tuba?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23639 + }, + { + "path": "esa00000021.wav", + "question": "Is the piano more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23640 + }, + { + "path": "esa00000640.wav", + "question": "Is the acoustic_guitar louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23641 + }, + { + "path": "esa00000182.wav", + "question": "Is the bagpipe more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "one", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23642 + }, + { + "path": "esa00000582.wav", + "question": "Is the electric_bass playing longer than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23643 + }, + { + "path": "esa00000463.wav", + "question": "Is the cello louder than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23644 + }, + { + "path": "sa00008641.wav", + "question": "Is the saxophone louder than the flute?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23645 + }, + { + "path": "eva00000853.wav", + "question": "Is the acoustic_guitar more rhythmic than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23646 + }, + { + "path": "esa00000354.wav", + "question": "Is the ukulele more rhythmic than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23647 + }, + { + "path": "esa00000222.wav", + "question": "Is the guzheng playing longer than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23648 + }, + { + "path": "esa00000269.wav", + "question": "Is the banjo playing longer than the trumpet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23649 + }, + { + "path": "evv00001467.wav", + "question": "Is the saxophone more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23650 + }, + { + "path": "eva00001011.wav", + "question": "Is the flute playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23651 + }, + { + "path": "00002957.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "one", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23652 + }, + { + "path": "esa00000007.wav", + "question": "Is the violin more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23653 + }, + { + "path": "esa00000210.wav", + "question": "Is the drum playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23654 + }, + { + "path": "eva00000940.wav", + "question": "Is the banjo more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23655 + }, + { + "path": "va00004887.wav", + "question": "Is the saxophone more rhythmic than the drum?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23656 + }, + { + "path": "esa00000418.wav", + "question": "Is the pipa more rhythmic than the guzheng?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23657 + }, + { + "path": "00000585.wav", + "question": "Is the trumpet more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23658 + }, + { + "path": "00003177.wav", + "question": "Is the guzheng more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23659 + }, + { + "path": "00006066.wav", + "question": "Is the erhu more rhythmic than the pipa?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23660 + }, + { + "path": "00003567.wav", + "question": "Is the piano playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23661 + }, + { + "path": "esa00000164.wav", + "question": "Is the guzheng playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23662 + }, + { + "path": "00002464.wav", + "question": "Is the piano playing longer than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23663 + }, + { + "path": "00008142.wav", + "question": "Is the drum more rhythmic than the xylophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "three", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23664 + }, + { + "path": "esa00000505.wav", + "question": "Is the piano more rhythmic than the clarinet?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23665 + }, + { + "path": "sa00007605.wav", + "question": "Is the drum more rhythmic than the electric_bass?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23666 + }, + { + "path": "va00005596.wav", + "question": "Is the clarinet more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "five", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23667 + }, + { + "path": "esa00000677.wav", + "question": "Is the xylophone playing longer than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23668 + }, + { + "path": "va00004770.wav", + "question": "Is the erhu playing longer than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "more than ten", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23669 + }, + { + "path": "sa00006515.wav", + "question": "Is the drum playing longer than the erhu?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23670 + }, + { + "path": "esa00000339.wav", + "question": "Is the guzheng playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "three", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23671 + }, + { + "path": "00003441.wav", + "question": "Is the flute playing longer than the xylophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23672 + }, + { + "path": "00007437.wav", + "question": "Is the piano playing longer than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "four", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23673 + }, + { + "path": "esa00000368.wav", + "question": "Is the saxophone more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23674 + }, + { + "path": "vv00002511.wav", + "question": "Is the violin playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23675 + }, + { + "path": "eva00001104.wav", + "question": "Is the piano playing longer than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "four", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23676 + }, + { + "path": "00003989.wav", + "question": "Is the piano more rhythmic than the electric_bass?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "more than ten", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23677 + }, + { + "path": "00003349.wav", + "question": "Is the saxophone more rhythmic than the acoustic_guitar?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "six", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23678 + }, + { + "path": "esa00000249.wav", + "question": "Is the congas playing longer than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "seven", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23679 + }, + { + "path": "00000957.wav", + "question": "Is the clarinet more rhythmic than the piano?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "six", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23680 + }, + { + "path": "esa00000688.wav", + "question": "Is the tuba more rhythmic than the violin?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "one", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23681 + }, + { + "path": "vv00000647.wav", + "question": "Is the suona more rhythmic than the saxophone?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "two", + "choice_d": "more than ten", + "answer_gt": "no", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23682 + }, + { + "path": "00003810.wav", + "question": "Is the clarinet playing longer than the bassoon?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "five", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23683 + }, + { + "path": "00005225.wav", + "question": "Is the ukulele more rhythmic than the banjo?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "seven", + "choice_d": "two", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23684 + }, + { + "path": "00007547.wav", + "question": "Is the violin more rhythmic than the cello?", + "choice_a": "yes", + "choice_b": "no", + "choice_c": "five", + "choice_d": "three", + "answer_gt": "yes", + "task_name": "Music_AQA", + "dataset_name": "music_avqa", + "uniq_id": 23685 + }, + { + "path": "23869.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "sport", + "choice_b": "positive", + "choice_c": "melodic", + "choice_d": "movie", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23686 + }, + { + "path": "1316280.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "advertising", + "choice_b": "happy", + "choice_c": "dark", + "choice_d": "relaxing", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23687 + }, + { + "path": "1370664.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "dark", + "choice_b": "slow", + "choice_c": "dream", + "choice_d": "funny", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23688 + }, + { + "path": "1280424.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "party", + "choice_b": "slow", + "choice_c": "adventure", + "choice_d": "dramatic", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23689 + }, + { + "path": "1379974.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "dream", + "choice_b": "game", + "choice_c": "melancholic", + "choice_d": "trailer", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23690 + }, + { + "path": "862197.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "action", + "choice_b": "energetic", + "choice_c": "deep", + "choice_d": "background", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23691 + }, + { + "path": "1420620.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "background", + "choice_b": "deep", + "choice_c": "sport", + "choice_d": "motivational", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23692 + }, + { + "path": "1350991.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "action", + "choice_b": "christmas", + "choice_c": "space", + "choice_d": "commercial", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23693 + }, + { + "path": "1159859.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "dream", + "choice_b": "space", + "choice_c": "drama", + "choice_d": "funny", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23694 + }, + { + "path": "1399624.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "powerful", + "choice_b": "summer", + "choice_c": "dramatic", + "choice_d": "soundscape", + "answer_gt": "dramatic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23695 + }, + { + "path": "1039171.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "summer", + "choice_b": "emotional", + "choice_c": "nature", + "choice_d": "travel", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23696 + }, + { + "path": "1415802.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "documentary", + "choice_b": "background", + "choice_c": "movie", + "choice_d": "slow", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23697 + }, + { + "path": "202648.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "deep", + "choice_b": "cool", + "choice_c": "advertising", + "choice_d": "film", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23698 + }, + { + "path": "1374316.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "upbeat", + "choice_b": "nature", + "choice_c": "game", + "choice_d": "dream", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23699 + }, + { + "path": "965303.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "movie", + "choice_b": "christmas", + "choice_c": "trailer", + "choice_d": "soft", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23700 + }, + { + "path": "1333716.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "commercial", + "choice_b": "corporate", + "choice_c": "dream", + "choice_d": "soundscape", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23701 + }, + { + "path": "979225.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "fun", + "choice_b": "background", + "choice_c": "melodic", + "choice_d": "advertising", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23702 + }, + { + "path": "965442.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "emotional", + "choice_b": "fun", + "choice_c": "christmas", + "choice_d": "sexy", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23703 + }, + { + "path": "663851.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "ballad", + "choice_b": "melodic", + "choice_c": "melancholic", + "choice_d": "game", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23704 + }, + { + "path": "1148029.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "trailer", + "choice_b": "film", + "choice_c": "soft", + "choice_d": "space", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23705 + }, + { + "path": "1349096.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "summer", + "choice_b": "heavy", + "choice_c": "sport", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23706 + }, + { + "path": "1186845.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "sad", + "choice_b": "motivational", + "choice_c": "emotional", + "choice_d": "upbeat", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23707 + }, + { + "path": "549228.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "ballad", + "choice_b": "holiday", + "choice_c": "energetic", + "choice_d": "upbeat", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23708 + }, + { + "path": "128248.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "christmas", + "choice_b": "energetic", + "choice_c": "deep", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23709 + }, + { + "path": "1319505.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "christmas", + "choice_b": "upbeat", + "choice_c": "sexy", + "choice_d": "soft", + "answer_gt": "sexy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23710 + }, + { + "path": "1312109.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "love", + "choice_b": "travel", + "choice_c": "movie", + "choice_d": "space", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23711 + }, + { + "path": "1350953.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "motivational", + "choice_b": "children", + "choice_c": "powerful", + "choice_d": "drama", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23712 + }, + { + "path": "1320589.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "powerful", + "choice_b": "groovy", + "choice_c": "upbeat", + "choice_d": "love", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23713 + }, + { + "path": "1314853.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "sexy", + "choice_b": "ballad", + "choice_c": "love", + "choice_d": "energetic", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23714 + }, + { + "path": "1134761.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "game", + "choice_b": "children", + "choice_c": "soundscape", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23715 + }, + { + "path": "1292119.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "dark", + "choice_b": "christmas", + "choice_c": "summer", + "choice_d": "soundscape", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23716 + }, + { + "path": "1148035.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "christmas", + "choice_b": "film", + "choice_c": "documentary", + "choice_d": "children", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23717 + }, + { + "path": "1348322.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "christmas", + "choice_b": "fast", + "choice_c": "melodic", + "choice_d": "game", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23718 + }, + { + "path": "859971.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "hopeful", + "choice_b": "soundscape", + "choice_c": "sport", + "choice_d": "space", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23719 + }, + { + "path": "85785.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "sexy", + "choice_b": "documentary", + "choice_c": "melodic", + "choice_d": "sport", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23720 + }, + { + "path": "1134327.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "children", + "choice_b": "soundscape", + "choice_c": "ballad", + "choice_d": "calm", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23721 + }, + { + "path": "1311818.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "fun", + "choice_b": "melancholic", + "choice_c": "romantic", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23722 + }, + { + "path": "1395611.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "sad", + "choice_b": "game", + "choice_c": "romantic", + "choice_d": "inspiring", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23723 + }, + { + "path": "1203657.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "summer", + "choice_b": "energetic", + "choice_c": "drama", + "choice_d": "fun", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23724 + }, + { + "path": "1420594.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "summer", + "choice_b": "deep", + "choice_c": "nature", + "choice_d": "christmas", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23725 + }, + { + "path": "467927.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "adventure", + "choice_b": "funny", + "choice_c": "film", + "choice_d": "dream", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23726 + }, + { + "path": "21861.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "advertising", + "choice_b": "heavy", + "choice_c": "christmas", + "choice_d": "background", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23727 + }, + { + "path": "1315325.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "deep", + "choice_b": "dramatic", + "choice_c": "party", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23728 + }, + { + "path": "1302993.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "christmas", + "choice_b": "positive", + "choice_c": "adventure", + "choice_d": "fast", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23729 + }, + { + "path": "85767.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "corporate", + "choice_b": "christmas", + "choice_c": "melodic", + "choice_d": "motivational", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23730 + }, + { + "path": "1330558.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "meditative", + "choice_b": "positive", + "choice_c": "trailer", + "choice_d": "advertising", + "answer_gt": "positive", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23731 + }, + { + "path": "1136983.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "corporate", + "choice_b": "relaxing", + "choice_c": "movie", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23732 + }, + { + "path": "1280428.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "inspiring", + "choice_b": "fast", + "choice_c": "party", + "choice_d": "background", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23733 + }, + { + "path": "1121567.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "melodic", + "choice_b": "heavy", + "choice_c": "relaxing", + "choice_d": "romantic", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23734 + }, + { + "path": "1050928.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "energetic", + "choice_b": "meditative", + "choice_c": "travel", + "choice_d": "deep", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23735 + }, + { + "path": "555370.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "party", + "choice_b": "ballad", + "choice_c": "soundscape", + "choice_d": "groovy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23736 + }, + { + "path": "242952.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "emotional", + "choice_b": "drama", + "choice_c": "love", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23737 + }, + { + "path": "965312.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "christmas", + "choice_b": "slow", + "choice_c": "motivational", + "choice_d": "corporate", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23738 + }, + { + "path": "1319500.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "corporate", + "choice_b": "nature", + "choice_c": "background", + "choice_d": "groovy", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23739 + }, + { + "path": "1353352.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "trailer", + "choice_b": "sexy", + "choice_c": "summer", + "choice_d": "fast", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23740 + }, + { + "path": "1085585.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "upbeat", + "choice_b": "meditative", + "choice_c": "party", + "choice_d": "relaxing", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23741 + }, + { + "path": "1053506.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "upbeat", + "choice_b": "deep", + "choice_c": "trailer", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23742 + }, + { + "path": "747442.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "soundscape", + "choice_b": "cool", + "choice_c": "heavy", + "choice_d": "romantic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23743 + }, + { + "path": "317724.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "film", + "choice_b": "upbeat", + "choice_c": "uplifting", + "choice_d": "happy", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23744 + }, + { + "path": "1419758.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "sport", + "choice_b": "slow", + "choice_c": "meditative", + "choice_d": "emotional", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23745 + }, + { + "path": "433645.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "advertising", + "choice_b": "children", + "choice_c": "soundscape", + "choice_d": "holiday", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23746 + }, + { + "path": "1362289.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "summer", + "choice_b": "happy", + "choice_c": "upbeat", + "choice_d": "soundscape", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23747 + }, + { + "path": "967573.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "heavy", + "choice_b": "holiday", + "choice_c": "dramatic", + "choice_d": "drama", + "answer_gt": "dramatic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23748 + }, + { + "path": "1314124.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "travel", + "choice_b": "deep", + "choice_c": "space", + "choice_d": "children", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23749 + }, + { + "path": "1286494.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "positive", + "choice_b": "dream", + "choice_c": "holiday", + "choice_d": "romantic", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23750 + }, + { + "path": "799923.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "travel", + "choice_b": "adventure", + "choice_c": "sport", + "choice_d": "ballad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23751 + }, + { + "path": "1056409.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "powerful", + "choice_b": "party", + "choice_c": "dream", + "choice_d": "epic", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23752 + }, + { + "path": "965438.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "christmas", + "choice_b": "relaxing", + "choice_c": "epic", + "choice_d": "corporate", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23753 + }, + { + "path": "1276495.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "soundscape", + "choice_b": "inspiring", + "choice_c": "game", + "choice_d": "movie", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23754 + }, + { + "path": "866706.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "dramatic", + "choice_b": "movie", + "choice_c": "film", + "choice_d": "christmas", + "answer_gt": "movie", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23755 + }, + { + "path": "1302904.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "soundscape", + "choice_b": "space", + "choice_c": "party", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23756 + }, + { + "path": "953706.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "emotional", + "choice_b": "movie", + "choice_c": "positive", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23757 + }, + { + "path": "231891.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "dream", + "choice_b": "slow", + "choice_c": "corporate", + "choice_d": "drama", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23758 + }, + { + "path": "1192523.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "cool", + "choice_b": "calm", + "choice_c": "documentary", + "choice_d": "commercial", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23759 + }, + { + "path": "635244.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "adventure", + "choice_b": "inspiring", + "choice_c": "dream", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23760 + }, + { + "path": "257173.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "meditative", + "choice_b": "film", + "choice_c": "hopeful", + "choice_d": "commercial", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23761 + }, + { + "path": "1103993.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "calm", + "choice_b": "ballad", + "choice_c": "christmas", + "choice_d": "romantic", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23762 + }, + { + "path": "146605.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "emotional", + "choice_b": "commercial", + "choice_c": "travel", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23763 + }, + { + "path": "1006558.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "soundscape", + "choice_b": "calm", + "choice_c": "action", + "choice_d": "dream", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23764 + }, + { + "path": "1199622.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "ballad", + "choice_b": "action", + "choice_c": "love", + "choice_d": "sport", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23765 + }, + { + "path": "73080.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "space", + "choice_b": "action", + "choice_c": "funny", + "choice_d": "relaxing", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23766 + }, + { + "path": "1168479.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "heavy", + "choice_b": "fast", + "choice_c": "christmas", + "choice_d": "inspiring", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23767 + }, + { + "path": "1420626.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "happy", + "choice_b": "nature", + "choice_c": "deep", + "choice_d": "meditative", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23768 + }, + { + "path": "1400517.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "melancholic", + "choice_b": "relaxing", + "choice_c": "slow", + "choice_d": "retro", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23769 + }, + { + "path": "1351054.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "background", + "choice_b": "children", + "choice_c": "game", + "choice_d": "hopeful", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23770 + }, + { + "path": "1203871.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "film", + "choice_b": "happy", + "choice_c": "meditative", + "choice_d": "adventure", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23771 + }, + { + "path": "1222032.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "travel", + "choice_b": "meditative", + "choice_c": "powerful", + "choice_d": "fun", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23772 + }, + { + "path": "1135069.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "melodic", + "choice_b": "emotional", + "choice_c": "summer", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23773 + }, + { + "path": "1160708.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "calm", + "choice_b": "upbeat", + "choice_c": "emotional", + "choice_d": "space", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23774 + }, + { + "path": "243435.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "funny", + "choice_b": "children", + "choice_c": "melodic", + "choice_d": "travel", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23775 + }, + { + "path": "1415944.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "groovy", + "choice_b": "trailer", + "choice_c": "movie", + "choice_d": "soft", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23776 + }, + { + "path": "1353350.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "epic", + "choice_b": "cool", + "choice_c": "commercial", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23777 + }, + { + "path": "1306958.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "sexy", + "choice_b": "movie", + "choice_c": "advertising", + "choice_d": "commercial", + "answer_gt": "movie", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23778 + }, + { + "path": "1131179.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "background", + "choice_b": "dream", + "choice_c": "nature", + "choice_d": "soundscape", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23779 + }, + { + "path": "175423.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "action", + "choice_b": "adventure", + "choice_c": "calm", + "choice_d": "children", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23780 + }, + { + "path": "1257052.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "ballad", + "choice_b": "holiday", + "choice_c": "nature", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23781 + }, + { + "path": "1232472.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "calm", + "choice_b": "happy", + "choice_c": "holiday", + "choice_d": "positive", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23782 + }, + { + "path": "1292127.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "dramatic", + "choice_b": "christmas", + "choice_c": "calm", + "choice_d": "groovy", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23783 + }, + { + "path": "1352367.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "energetic", + "choice_b": "nature", + "choice_c": "heavy", + "choice_d": "drama", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23784 + }, + { + "path": "1280922.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "summer", + "choice_b": "travel", + "choice_c": "dramatic", + "choice_d": "melodic", + "answer_gt": "dramatic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23785 + }, + { + "path": "949028.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "retro", + "choice_b": "melancholic", + "choice_c": "upbeat", + "choice_d": "adventure", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23786 + }, + { + "path": "1380711.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "funny", + "choice_b": "happy", + "choice_c": "sexy", + "choice_d": "fast", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23787 + }, + { + "path": "883576.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "deep", + "choice_b": "action", + "choice_c": "epic", + "choice_d": "sad", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23788 + }, + { + "path": "577043.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "fun", + "choice_b": "energetic", + "choice_c": "emotional", + "choice_d": "christmas", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23789 + }, + { + "path": "1199626.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "love", + "choice_b": "ballad", + "choice_c": "dark", + "choice_d": "epic", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23790 + }, + { + "path": "1193445.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "dramatic", + "choice_b": "happy", + "choice_c": "melodic", + "choice_d": "drama", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23791 + }, + { + "path": "973752.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "uplifting", + "choice_b": "christmas", + "choice_c": "film", + "choice_d": "motivational", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23792 + }, + { + "path": "859972.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "documentary", + "choice_b": "holiday", + "choice_c": "soundscape", + "choice_d": "children", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23793 + }, + { + "path": "1353345.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "positive", + "choice_b": "space", + "choice_c": "summer", + "choice_d": "fast", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23794 + }, + { + "path": "1131177.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "christmas", + "choice_b": "dramatic", + "choice_c": "fast", + "choice_d": "nature", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23795 + }, + { + "path": "1274423.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "meditative", + "choice_b": "groovy", + "choice_c": "sexy", + "choice_d": "holiday", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23796 + }, + { + "path": "1311076.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "film", + "choice_b": "motivational", + "choice_c": "sport", + "choice_d": "funny", + "answer_gt": "motivational", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23797 + }, + { + "path": "1160601.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "powerful", + "choice_b": "christmas", + "choice_c": "melancholic", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23798 + }, + { + "path": "1321603.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "melodic", + "choice_b": "holiday", + "choice_c": "positive", + "choice_d": "fun", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23799 + }, + { + "path": "1291246.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "drama", + "choice_b": "motivational", + "choice_c": "uplifting", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23800 + }, + { + "path": "1056411.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "drama", + "choice_b": "adventure", + "choice_c": "heavy", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23801 + }, + { + "path": "1145441.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "motivational", + "choice_b": "dark", + "choice_c": "melancholic", + "choice_d": "drama", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23802 + }, + { + "path": "494291.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "soft", + "choice_b": "dark", + "choice_c": "relaxing", + "choice_d": "nature", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23803 + }, + { + "path": "1294935.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "deep", + "choice_b": "trailer", + "choice_c": "christmas", + "choice_d": "melancholic", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23804 + }, + { + "path": "1371514.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "space", + "choice_b": "positive", + "choice_c": "summer", + "choice_d": "slow", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23805 + }, + { + "path": "1318026.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "energetic", + "choice_b": "dream", + "choice_c": "epic", + "choice_d": "hopeful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23806 + }, + { + "path": "1420605.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "deep", + "choice_b": "corporate", + "choice_c": "upbeat", + "choice_d": "dramatic", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23807 + }, + { + "path": "283809.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "ballad", + "choice_b": "drama", + "choice_c": "meditative", + "choice_d": "documentary", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23808 + }, + { + "path": "1273972.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "ballad", + "choice_b": "dramatic", + "choice_c": "epic", + "choice_d": "advertising", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23809 + }, + { + "path": "6608.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "powerful", + "choice_b": "melodic", + "choice_c": "action", + "choice_d": "upbeat", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23810 + }, + { + "path": "828836.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "game", + "choice_b": "space", + "choice_c": "uplifting", + "choice_d": "dark", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23811 + }, + { + "path": "1352368.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "energetic", + "choice_b": "sport", + "choice_c": "soft", + "choice_d": "adventure", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23812 + }, + { + "path": "175669.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "dream", + "choice_b": "energetic", + "choice_c": "dramatic", + "choice_d": "trailer", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23813 + }, + { + "path": "285042.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "movie", + "choice_b": "upbeat", + "choice_c": "retro", + "choice_d": "dream", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23814 + }, + { + "path": "1313336.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "energetic", + "choice_b": "inspiring", + "choice_c": "dramatic", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23815 + }, + { + "path": "1192554.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "dream", + "choice_b": "christmas", + "choice_c": "emotional", + "choice_d": "action", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23816 + }, + { + "path": "1353452.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "holiday", + "choice_b": "christmas", + "choice_c": "sexy", + "choice_d": "sad", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23817 + }, + { + "path": "923464.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "travel", + "choice_b": "film", + "choice_c": "fun", + "choice_d": "groovy", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23818 + }, + { + "path": "1158172.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "fast", + "choice_b": "energetic", + "choice_c": "positive", + "choice_d": "sad", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23819 + }, + { + "path": "287786.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "funny", + "choice_b": "deep", + "choice_c": "energetic", + "choice_d": "children", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23820 + }, + { + "path": "1319330.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "corporate", + "choice_b": "dark", + "choice_c": "melodic", + "choice_d": "inspiring", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23821 + }, + { + "path": "1361471.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "uplifting", + "choice_b": "commercial", + "choice_c": "nature", + "choice_d": "sad", + "answer_gt": "commercial", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23822 + }, + { + "path": "1096867.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "dramatic", + "choice_b": "motivational", + "choice_c": "game", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23823 + }, + { + "path": "1160603.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "children", + "choice_b": "movie", + "choice_c": "melodic", + "choice_d": "corporate", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23824 + }, + { + "path": "856979.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "energetic", + "choice_b": "fun", + "choice_c": "dark", + "choice_d": "heavy", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23825 + }, + { + "path": "1161202.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "commercial", + "choice_b": "space", + "choice_c": "upbeat", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23826 + }, + { + "path": "1121761.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "hopeful", + "choice_b": "happy", + "choice_c": "meditative", + "choice_d": "sport", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23827 + }, + { + "path": "433648.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "groovy", + "choice_b": "soundscape", + "choice_c": "ballad", + "choice_d": "children", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23828 + }, + { + "path": "283091.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "relaxing", + "choice_b": "corporate", + "choice_c": "commercial", + "choice_d": "sport", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23829 + }, + { + "path": "549221.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "ballad", + "choice_b": "documentary", + "choice_c": "game", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23830 + }, + { + "path": "1318025.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "dream", + "choice_b": "slow", + "choice_c": "documentary", + "choice_d": "powerful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23831 + }, + { + "path": "1394753.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "emotional", + "choice_b": "melancholic", + "choice_c": "ballad", + "choice_d": "uplifting", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23832 + }, + { + "path": "1045835.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "background", + "choice_b": "energetic", + "choice_c": "space", + "choice_d": "retro", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23833 + }, + { + "path": "1256654.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "dark", + "choice_b": "sexy", + "choice_c": "space", + "choice_d": "party", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23834 + }, + { + "path": "23866.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "melodic", + "choice_b": "uplifting", + "choice_c": "soundscape", + "choice_d": "nature", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23835 + }, + { + "path": "231492.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "christmas", + "choice_b": "summer", + "choice_c": "party", + "choice_d": "film", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23836 + }, + { + "path": "317048.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "fast", + "choice_b": "corporate", + "choice_c": "hopeful", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23837 + }, + { + "path": "1168674.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "ballad", + "choice_b": "emotional", + "choice_c": "travel", + "choice_d": "calm", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23838 + }, + { + "path": "17156.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "dream", + "choice_b": "melodic", + "choice_c": "space", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23839 + }, + { + "path": "1370943.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "documentary", + "choice_b": "film", + "choice_c": "cool", + "choice_d": "background", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23840 + }, + { + "path": "384141.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "movie", + "choice_b": "upbeat", + "choice_c": "energetic", + "choice_d": "hopeful", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23841 + }, + { + "path": "1148033.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "film", + "choice_b": "commercial", + "choice_c": "groovy", + "choice_d": "positive", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23842 + }, + { + "path": "1095365.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "groovy", + "choice_b": "emotional", + "choice_c": "sport", + "choice_d": "deep", + "answer_gt": "sport", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23843 + }, + { + "path": "1327467.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "sport", + "choice_b": "sad", + "choice_c": "hopeful", + "choice_d": "party", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23844 + }, + { + "path": "1301819.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "relaxing", + "choice_b": "sexy", + "choice_c": "emotional", + "choice_d": "party", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23845 + }, + { + "path": "1294941.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "melodic", + "choice_b": "relaxing", + "choice_c": "nature", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23846 + }, + { + "path": "132207.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "dream", + "choice_b": "sport", + "choice_c": "cool", + "choice_d": "powerful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23847 + }, + { + "path": "25227.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "inspiring", + "choice_b": "melancholic", + "choice_c": "positive", + "choice_d": "dream", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23848 + }, + { + "path": "114200.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "game", + "choice_b": "meditative", + "choice_c": "upbeat", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23849 + }, + { + "path": "1102300.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "space", + "choice_b": "summer", + "choice_c": "melodic", + "choice_d": "fast", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23850 + }, + { + "path": "1349122.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "documentary", + "choice_b": "emotional", + "choice_c": "powerful", + "choice_d": "positive", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23851 + }, + { + "path": "1356320.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "corporate", + "choice_b": "holiday", + "choice_c": "game", + "choice_d": "relaxing", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23852 + }, + { + "path": "208940.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "energetic", + "choice_b": "nature", + "choice_c": "melodic", + "choice_d": "romantic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23853 + }, + { + "path": "1078363.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "retro", + "choice_b": "dream", + "choice_c": "christmas", + "choice_d": "inspiring", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23854 + }, + { + "path": "419240.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "melodic", + "choice_b": "sexy", + "choice_c": "corporate", + "choice_d": "melancholic", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23855 + }, + { + "path": "784526.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "holiday", + "choice_b": "dream", + "choice_c": "deep", + "choice_d": "calm", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23856 + }, + { + "path": "1204594.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "film", + "choice_b": "background", + "choice_c": "ballad", + "choice_d": "party", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23857 + }, + { + "path": "23861.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "holiday", + "choice_b": "game", + "choice_c": "drama", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23858 + }, + { + "path": "76989.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "commercial", + "choice_b": "dark", + "choice_c": "energetic", + "choice_d": "space", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23859 + }, + { + "path": "1286513.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "romantic", + "choice_b": "happy", + "choice_c": "space", + "choice_d": "powerful", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23860 + }, + { + "path": "290863.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "adventure", + "choice_b": "sad", + "choice_c": "cool", + "choice_d": "soundscape", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23861 + }, + { + "path": "1353451.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "powerful", + "choice_b": "action", + "choice_c": "christmas", + "choice_d": "melodic", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23862 + }, + { + "path": "1131999.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "adventure", + "choice_b": "christmas", + "choice_c": "action", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23863 + }, + { + "path": "1276174.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "background", + "choice_b": "game", + "choice_c": "film", + "choice_d": "inspiring", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23864 + }, + { + "path": "80374.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "christmas", + "choice_b": "melancholic", + "choice_c": "documentary", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23865 + }, + { + "path": "1377209.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "positive", + "choice_b": "groovy", + "choice_c": "upbeat", + "choice_d": "melodic", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23866 + }, + { + "path": "1049828.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "movie", + "choice_b": "energetic", + "choice_c": "children", + "choice_d": "retro", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23867 + }, + { + "path": "1235974.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "soundscape", + "choice_b": "upbeat", + "choice_c": "emotional", + "choice_d": "romantic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23868 + }, + { + "path": "1049833.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "energetic", + "choice_b": "sport", + "choice_c": "background", + "choice_d": "christmas", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23869 + }, + { + "path": "14021.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "fast", + "choice_b": "slow", + "choice_c": "calm", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23870 + }, + { + "path": "127888.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "trailer", + "choice_b": "calm", + "choice_c": "dream", + "choice_d": "melodic", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23871 + }, + { + "path": "686360.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "game", + "choice_b": "relaxing", + "choice_c": "melodic", + "choice_d": "groovy", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23872 + }, + { + "path": "1394402.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "christmas", + "choice_b": "powerful", + "choice_c": "dark", + "choice_d": "game", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23873 + }, + { + "path": "1136364.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "soundscape", + "choice_b": "upbeat", + "choice_c": "summer", + "choice_d": "movie", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23874 + }, + { + "path": "918148.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "soft", + "choice_b": "powerful", + "choice_c": "positive", + "choice_d": "travel", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23875 + }, + { + "path": "1245889.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "slow", + "choice_b": "groovy", + "choice_c": "movie", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23876 + }, + { + "path": "1279452.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "dream", + "choice_b": "christmas", + "choice_c": "deep", + "choice_d": "space", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23877 + }, + { + "path": "941853.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "adventure", + "choice_b": "funny", + "choice_c": "dream", + "choice_d": "groovy", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23878 + }, + { + "path": "1350819.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "funny", + "choice_b": "children", + "choice_c": "holiday", + "choice_d": "emotional", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23879 + }, + { + "path": "1116453.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "slow", + "choice_b": "film", + "choice_c": "ballad", + "choice_d": "fun", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23880 + }, + { + "path": "1221466.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "love", + "choice_b": "action", + "choice_c": "uplifting", + "choice_d": "powerful", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23881 + }, + { + "path": "1311064.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "motivational", + "choice_b": "deep", + "choice_c": "holiday", + "choice_d": "funny", + "answer_gt": "motivational", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23882 + }, + { + "path": "597313.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "holiday", + "choice_b": "slow", + "choice_c": "dark", + "choice_d": "fast", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23883 + }, + { + "path": "1420602.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "deep", + "choice_b": "energetic", + "choice_c": "upbeat", + "choice_d": "sport", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23884 + }, + { + "path": "1078365.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "travel", + "choice_b": "adventure", + "choice_c": "christmas", + "choice_d": "holiday", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23885 + }, + { + "path": "701791.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "christmas", + "choice_b": "action", + "choice_c": "romantic", + "choice_d": "film", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23886 + }, + { + "path": "1121487.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "melancholic", + "choice_b": "dream", + "choice_c": "summer", + "choice_d": "adventure", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23887 + }, + { + "path": "726908.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "dream", + "choice_b": "summer", + "choice_c": "adventure", + "choice_d": "groovy", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23888 + }, + { + "path": "257205.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "relaxing", + "choice_b": "ballad", + "choice_c": "deep", + "choice_d": "epic", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23889 + }, + { + "path": "1220668.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "epic", + "choice_b": "party", + "choice_c": "fast", + "choice_d": "calm", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23890 + }, + { + "path": "33825.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "documentary", + "choice_b": "children", + "choice_c": "powerful", + "choice_d": "advertising", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23891 + }, + { + "path": "1353781.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "positive", + "choice_b": "funny", + "choice_c": "commercial", + "choice_d": "sexy", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23892 + }, + { + "path": "869207.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "relaxing", + "choice_b": "dream", + "choice_c": "heavy", + "choice_d": "melancholic", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23893 + }, + { + "path": "1123936.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "fun", + "choice_b": "emotional", + "choice_c": "travel", + "choice_d": "christmas", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23894 + }, + { + "path": "1273388.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "travel", + "choice_b": "film", + "choice_c": "melodic", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23895 + }, + { + "path": "264448.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "hopeful", + "choice_b": "soundscape", + "choice_c": "sad", + "choice_d": "sexy", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23896 + }, + { + "path": "1056231.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "adventure", + "choice_b": "summer", + "choice_c": "dramatic", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23897 + }, + { + "path": "875380.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "dramatic", + "choice_b": "soundscape", + "choice_c": "summer", + "choice_d": "fast", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23898 + }, + { + "path": "865772.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "drama", + "choice_b": "advertising", + "choice_c": "slow", + "choice_d": "positive", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23899 + }, + { + "path": "579302.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "dramatic", + "choice_b": "movie", + "choice_c": "upbeat", + "choice_d": "action", + "answer_gt": "action", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23900 + }, + { + "path": "973757.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "corporate", + "choice_b": "children", + "choice_c": "energetic", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23901 + }, + { + "path": "911677.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "film", + "choice_b": "motivational", + "choice_c": "deep", + "choice_d": "holiday", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23902 + }, + { + "path": "1163350.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "film", + "choice_b": "fun", + "choice_c": "children", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23903 + }, + { + "path": "457515.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "energetic", + "choice_b": "uplifting", + "choice_c": "nature", + "choice_d": "love", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23904 + }, + { + "path": "1094520.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "nature", + "choice_b": "retro", + "choice_c": "background", + "choice_d": "romantic", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23905 + }, + { + "path": "1292125.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "melodic", + "choice_b": "emotional", + "choice_c": "dream", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23906 + }, + { + "path": "862194.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "sad", + "choice_b": "inspiring", + "choice_c": "deep", + "choice_d": "groovy", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23907 + }, + { + "path": "824632.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "dark", + "choice_b": "fun", + "choice_c": "meditative", + "choice_d": "film", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23908 + }, + { + "path": "1167071.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "advertising", + "choice_b": "game", + "choice_c": "upbeat", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23909 + }, + { + "path": "17152.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "background", + "choice_b": "nature", + "choice_c": "dark", + "choice_d": "summer", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23910 + }, + { + "path": "1192534.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "emotional", + "choice_b": "documentary", + "choice_c": "movie", + "choice_d": "dramatic", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23911 + }, + { + "path": "433640.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "children", + "choice_b": "groovy", + "choice_c": "soundscape", + "choice_d": "sport", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23912 + }, + { + "path": "1063678.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "space", + "choice_b": "film", + "choice_c": "sport", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23913 + }, + { + "path": "356420.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "drama", + "choice_b": "ballad", + "choice_c": "inspiring", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23914 + }, + { + "path": "921856.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "emotional", + "choice_b": "motivational", + "choice_c": "dark", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23915 + }, + { + "path": "1115570.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "sport", + "choice_b": "energetic", + "choice_c": "summer", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23916 + }, + { + "path": "263535.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "happy", + "choice_b": "sport", + "choice_c": "romantic", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23917 + }, + { + "path": "493295.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "fast", + "choice_b": "children", + "choice_c": "travel", + "choice_d": "holiday", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23918 + }, + { + "path": "1124712.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "dark", + "choice_b": "energetic", + "choice_c": "hopeful", + "choice_d": "happy", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23919 + }, + { + "path": "840600.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "soft", + "choice_b": "happy", + "choice_c": "summer", + "choice_d": "love", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23920 + }, + { + "path": "46195.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "calm", + "choice_b": "hopeful", + "choice_c": "documentary", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23921 + }, + { + "path": "916394.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "energetic", + "choice_b": "upbeat", + "choice_c": "fun", + "choice_d": "dark", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23922 + }, + { + "path": "710694.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "dark", + "choice_b": "hopeful", + "choice_c": "melodic", + "choice_d": "ballad", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23923 + }, + { + "path": "1280425.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "sad", + "choice_b": "advertising", + "choice_c": "soft", + "choice_d": "party", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23924 + }, + { + "path": "1221463.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "cool", + "choice_b": "action", + "choice_c": "sport", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23925 + }, + { + "path": "1089272.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "dramatic", + "choice_b": "sexy", + "choice_c": "game", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23926 + }, + { + "path": "620609.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "energetic", + "choice_b": "soft", + "choice_c": "cool", + "choice_d": "game", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23927 + }, + { + "path": "1118532.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "travel", + "choice_b": "romantic", + "choice_c": "upbeat", + "choice_d": "space", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23928 + }, + { + "path": "1420599.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "slow", + "choice_b": "motivational", + "choice_c": "corporate", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23929 + }, + { + "path": "1053503.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "energetic", + "choice_b": "inspiring", + "choice_c": "trailer", + "choice_d": "children", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23930 + }, + { + "path": "1160234.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "corporate", + "choice_b": "calm", + "choice_c": "love", + "choice_d": "adventure", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23931 + }, + { + "path": "14019.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "romantic", + "choice_b": "advertising", + "choice_c": "relaxing", + "choice_d": "groovy", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23932 + }, + { + "path": "1226198.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "advertising", + "choice_b": "commercial", + "choice_c": "space", + "choice_d": "motivational", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23933 + }, + { + "path": "1381056.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "action", + "choice_b": "romantic", + "choice_c": "inspiring", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23934 + }, + { + "path": "25237.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "happy", + "choice_b": "movie", + "choice_c": "fun", + "choice_d": "space", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23935 + }, + { + "path": "872920.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "relaxing", + "choice_b": "positive", + "choice_c": "deep", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23936 + }, + { + "path": "1221326.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "retro", + "choice_b": "melodic", + "choice_c": "relaxing", + "choice_d": "meditative", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23937 + }, + { + "path": "1393139.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "christmas", + "choice_b": "advertising", + "choice_c": "emotional", + "choice_d": "dream", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23938 + }, + { + "path": "1179522.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "inspiring", + "choice_b": "uplifting", + "choice_c": "motivational", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23939 + }, + { + "path": "1229453.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "hopeful", + "choice_b": "game", + "choice_c": "groovy", + "choice_d": "movie", + "answer_gt": "movie", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23940 + }, + { + "path": "378474.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "commercial", + "choice_b": "soundscape", + "choice_c": "fast", + "choice_d": "cool", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23941 + }, + { + "path": "840599.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "fast", + "choice_b": "happy", + "choice_c": "upbeat", + "choice_d": "dark", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23942 + }, + { + "path": "1095788.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "powerful", + "choice_b": "retro", + "choice_c": "motivational", + "choice_d": "sexy", + "answer_gt": "motivational", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23943 + }, + { + "path": "1350818.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "inspiring", + "choice_b": "documentary", + "choice_c": "children", + "choice_d": "heavy", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23944 + }, + { + "path": "1347954.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "retro", + "choice_b": "summer", + "choice_c": "dream", + "choice_d": "movie", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23945 + }, + { + "path": "1273387.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "slow", + "choice_b": "dramatic", + "choice_c": "fast", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23946 + }, + { + "path": "6610.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "melodic", + "choice_b": "sport", + "choice_c": "trailer", + "choice_d": "positive", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23947 + }, + { + "path": "1094474.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "advertising", + "choice_b": "space", + "choice_c": "energetic", + "choice_d": "funny", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23948 + }, + { + "path": "1280363.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "calm", + "choice_b": "soft", + "choice_c": "uplifting", + "choice_d": "funny", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23949 + }, + { + "path": "376420.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "deep", + "choice_b": "love", + "choice_c": "holiday", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23950 + }, + { + "path": "1277276.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "soundscape", + "choice_b": "sad", + "choice_c": "funny", + "choice_d": "energetic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23951 + }, + { + "path": "1012513.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "melodic", + "choice_b": "positive", + "choice_c": "space", + "choice_d": "energetic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23952 + }, + { + "path": "1257049.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "melodic", + "choice_b": "calm", + "choice_c": "children", + "choice_d": "documentary", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23953 + }, + { + "path": "1420628.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "deep", + "choice_b": "party", + "choice_c": "inspiring", + "choice_d": "sport", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23954 + }, + { + "path": "965446.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "christmas", + "choice_b": "corporate", + "choice_c": "positive", + "choice_d": "fun", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23955 + }, + { + "path": "1131174.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "children", + "choice_b": "heavy", + "choice_c": "relaxing", + "choice_d": "nature", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23956 + }, + { + "path": "1089275.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "powerful", + "choice_b": "upbeat", + "choice_c": "ballad", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23957 + }, + { + "path": "894531.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "dark", + "choice_b": "party", + "choice_c": "inspiring", + "choice_d": "sport", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23958 + }, + { + "path": "1170739.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "children", + "choice_b": "holiday", + "choice_c": "relaxing", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23959 + }, + { + "path": "1243719.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "soundscape", + "choice_b": "happy", + "choice_c": "heavy", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23960 + }, + { + "path": "1244453.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "cool", + "choice_b": "motivational", + "choice_c": "inspiring", + "choice_d": "trailer", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23961 + }, + { + "path": "1068198.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "energetic", + "choice_b": "travel", + "choice_c": "meditative", + "choice_d": "dream", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23962 + }, + { + "path": "1273390.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "ballad", + "choice_b": "dark", + "choice_c": "happy", + "choice_d": "retro", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23963 + }, + { + "path": "965307.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "soundscape", + "choice_b": "deep", + "choice_c": "groovy", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23964 + }, + { + "path": "114201.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "energetic", + "choice_b": "calm", + "choice_c": "relaxing", + "choice_d": "documentary", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23965 + }, + { + "path": "979224.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "fun", + "choice_b": "space", + "choice_c": "inspiring", + "choice_d": "dark", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23966 + }, + { + "path": "1097804.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "dark", + "choice_b": "melodic", + "choice_c": "film", + "choice_d": "upbeat", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23967 + }, + { + "path": "828837.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "space", + "choice_b": "soundscape", + "choice_c": "slow", + "choice_d": "happy", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23968 + }, + { + "path": "1104259.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "deep", + "choice_b": "inspiring", + "choice_c": "corporate", + "choice_d": "trailer", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23969 + }, + { + "path": "1170731.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "christmas", + "choice_b": "fast", + "choice_c": "action", + "choice_d": "love", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23970 + }, + { + "path": "1248617.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "love", + "choice_b": "space", + "choice_c": "dark", + "choice_d": "party", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23971 + }, + { + "path": "1198577.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "adventure", + "choice_b": "energetic", + "choice_c": "soft", + "choice_d": "heavy", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23972 + }, + { + "path": "1221325.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "happy", + "choice_b": "action", + "choice_c": "dream", + "choice_d": "dark", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23973 + }, + { + "path": "973755.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "motivational", + "choice_b": "christmas", + "choice_c": "inspiring", + "choice_d": "party", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23974 + }, + { + "path": "283097.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "melancholic", + "choice_b": "party", + "choice_c": "heavy", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23975 + }, + { + "path": "1353347.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "relaxing", + "choice_b": "meditative", + "choice_c": "advertising", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23976 + }, + { + "path": "1155058.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "energetic", + "choice_b": "background", + "choice_c": "space", + "choice_d": "heavy", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23977 + }, + { + "path": "1309626.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "trailer", + "choice_b": "energetic", + "choice_c": "inspiring", + "choice_d": "sad", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23978 + }, + { + "path": "786782.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "emotional", + "choice_b": "upbeat", + "choice_c": "melancholic", + "choice_d": "meditative", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23979 + }, + { + "path": "875376.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "space", + "choice_b": "fun", + "choice_c": "groovy", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23980 + }, + { + "path": "1241457.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "action", + "choice_b": "love", + "choice_c": "uplifting", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23981 + }, + { + "path": "1053507.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "advertising", + "choice_b": "retro", + "choice_c": "energetic", + "choice_d": "party", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23982 + }, + { + "path": "1225956.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "romantic", + "choice_b": "trailer", + "choice_c": "melodic", + "choice_d": "slow", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23983 + }, + { + "path": "208939.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "ballad", + "choice_b": "melodic", + "choice_c": "romantic", + "choice_d": "documentary", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23984 + }, + { + "path": "132217.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "fast", + "choice_b": "powerful", + "choice_c": "slow", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23985 + }, + { + "path": "884174.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "epic", + "choice_b": "game", + "choice_c": "background", + "choice_d": "inspiring", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23986 + }, + { + "path": "1226201.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "epic", + "choice_b": "space", + "choice_c": "ballad", + "choice_d": "hopeful", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23987 + }, + { + "path": "1186026.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "soft", + "choice_b": "energetic", + "choice_c": "drama", + "choice_d": "advertising", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23988 + }, + { + "path": "493297.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "motivational", + "choice_b": "dramatic", + "choice_c": "love", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23989 + }, + { + "path": "1134756.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "energetic", + "choice_b": "film", + "choice_c": "nature", + "choice_d": "soft", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23990 + }, + { + "path": "1041104.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "groovy", + "choice_b": "energetic", + "choice_c": "retro", + "choice_d": "calm", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23991 + }, + { + "path": "588709.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "melodic", + "choice_b": "retro", + "choice_c": "hopeful", + "choice_d": "motivational", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23992 + }, + { + "path": "1063677.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "summer", + "choice_b": "soundscape", + "choice_c": "motivational", + "choice_d": "retro", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23993 + }, + { + "path": "1365997.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "slow", + "choice_b": "sport", + "choice_c": "holiday", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23994 + }, + { + "path": "715130.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "party", + "choice_b": "cool", + "choice_c": "dramatic", + "choice_d": "holiday", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23995 + }, + { + "path": "1108513.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "background", + "choice_b": "hopeful", + "choice_c": "heavy", + "choice_d": "drama", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23996 + }, + { + "path": "467928.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "movie", + "choice_b": "corporate", + "choice_c": "travel", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23997 + }, + { + "path": "1280360.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "calm", + "choice_b": "meditative", + "choice_c": "hopeful", + "choice_d": "emotional", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23998 + }, + { + "path": "887209.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "soft", + "choice_b": "energetic", + "choice_c": "nature", + "choice_d": "calm", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 23999 + }, + { + "path": "1107277.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "groovy", + "choice_b": "epic", + "choice_c": "calm", + "choice_d": "sport", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24000 + }, + { + "path": "1312103.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "party", + "choice_b": "trailer", + "choice_c": "love", + "choice_d": "calm", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24001 + }, + { + "path": "1131181.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "nature", + "choice_b": "soundscape", + "choice_c": "deep", + "choice_d": "cool", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24002 + }, + { + "path": "1029840.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "space", + "choice_b": "holiday", + "choice_c": "documentary", + "choice_d": "sexy", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24003 + }, + { + "path": "1248601.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "party", + "choice_b": "trailer", + "choice_c": "dark", + "choice_d": "romantic", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24004 + }, + { + "path": "1278517.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "sexy", + "choice_b": "happy", + "choice_c": "epic", + "choice_d": "summer", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24005 + }, + { + "path": "1356317.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "dream", + "choice_b": "upbeat", + "choice_c": "motivational", + "choice_d": "cool", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24006 + }, + { + "path": "1003073.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "film", + "choice_b": "corporate", + "choice_c": "melodic", + "choice_d": "soft", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24007 + }, + { + "path": "1276525.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "dramatic", + "choice_b": "trailer", + "choice_c": "party", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24008 + }, + { + "path": "121123.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "travel", + "choice_b": "trailer", + "choice_c": "fast", + "choice_d": "documentary", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24009 + }, + { + "path": "1047960.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "dark", + "choice_b": "corporate", + "choice_c": "fast", + "choice_d": "dream", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24010 + }, + { + "path": "945916.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "fun", + "choice_b": "retro", + "choice_c": "dream", + "choice_d": "travel", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24011 + }, + { + "path": "1309614.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "retro", + "choice_b": "emotional", + "choice_c": "inspiring", + "choice_d": "epic", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24012 + }, + { + "path": "1314984.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "commercial", + "choice_b": "soundscape", + "choice_c": "positive", + "choice_d": "background", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24013 + }, + { + "path": "96720.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "dream", + "choice_b": "travel", + "choice_c": "funny", + "choice_d": "groovy", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24014 + }, + { + "path": "1163784.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "inspiring", + "choice_b": "game", + "choice_c": "retro", + "choice_d": "soft", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24015 + }, + { + "path": "1342137.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "happy", + "choice_b": "dramatic", + "choice_c": "commercial", + "choice_d": "motivational", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24016 + }, + { + "path": "365593.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "calm", + "choice_b": "emotional", + "choice_c": "groovy", + "choice_d": "background", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24017 + }, + { + "path": "1316286.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "fast", + "choice_b": "action", + "choice_c": "documentary", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24018 + }, + { + "path": "1109581.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "adventure", + "choice_b": "fun", + "choice_c": "inspiring", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24019 + }, + { + "path": "1100002.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "children", + "choice_b": "meditative", + "choice_c": "dream", + "choice_d": "movie", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24020 + }, + { + "path": "1312227.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "epic", + "choice_b": "soundscape", + "choice_c": "hopeful", + "choice_d": "sad", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24021 + }, + { + "path": "1108512.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "happy", + "choice_b": "background", + "choice_c": "summer", + "choice_d": "sport", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24022 + }, + { + "path": "973758.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "deep", + "choice_b": "fun", + "choice_c": "upbeat", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24023 + }, + { + "path": "1119164.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "hopeful", + "choice_b": "children", + "choice_c": "film", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24024 + }, + { + "path": "1419769.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "positive", + "choice_b": "retro", + "choice_c": "commercial", + "choice_d": "fast", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24025 + }, + { + "path": "1154295.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "sad", + "choice_b": "meditative", + "choice_c": "sport", + "choice_d": "slow", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24026 + }, + { + "path": "1303945.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "soundscape", + "choice_b": "melancholic", + "choice_c": "advertising", + "choice_d": "motivational", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24027 + }, + { + "path": "1370658.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "groovy", + "choice_b": "soft", + "choice_c": "relaxing", + "choice_d": "fun", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24028 + }, + { + "path": "1289386.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "drama", + "choice_b": "slow", + "choice_c": "action", + "choice_d": "soft", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24029 + }, + { + "path": "941857.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "dream", + "choice_b": "sexy", + "choice_c": "epic", + "choice_d": "relaxing", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24030 + }, + { + "path": "356418.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "heavy", + "choice_b": "dream", + "choice_c": "sad", + "choice_d": "party", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24031 + }, + { + "path": "1420534.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "documentary", + "choice_b": "drama", + "choice_c": "retro", + "choice_d": "soundscape", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24032 + }, + { + "path": "1349187.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "positive", + "choice_b": "energetic", + "choice_c": "dark", + "choice_d": "ballad", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24033 + }, + { + "path": "1031108.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "slow", + "choice_b": "advertising", + "choice_c": "epic", + "choice_d": "energetic", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24034 + }, + { + "path": "1236010.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "party", + "choice_b": "soundscape", + "choice_c": "corporate", + "choice_d": "ballad", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24035 + }, + { + "path": "1063679.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "energetic", + "choice_b": "commercial", + "choice_c": "soundscape", + "choice_d": "hopeful", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24036 + }, + { + "path": "620617.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "dark", + "choice_b": "emotional", + "choice_c": "adventure", + "choice_d": "ballad", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24037 + }, + { + "path": "114196.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "relaxing", + "choice_b": "cool", + "choice_c": "slow", + "choice_d": "dream", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24038 + }, + { + "path": "440767.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "fun", + "choice_b": "trailer", + "choice_c": "christmas", + "choice_d": "soundscape", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24039 + }, + { + "path": "1329121.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "meditative", + "choice_b": "hopeful", + "choice_c": "powerful", + "choice_d": "groovy", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24040 + }, + { + "path": "875382.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "soundscape", + "choice_b": "fast", + "choice_c": "background", + "choice_d": "funny", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24041 + }, + { + "path": "1149753.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "corporate", + "choice_b": "cool", + "choice_c": "inspiring", + "choice_d": "sad", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24042 + }, + { + "path": "1374324.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "groovy", + "choice_b": "drama", + "choice_c": "melodic", + "choice_d": "space", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24043 + }, + { + "path": "686356.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "romantic", + "choice_b": "relaxing", + "choice_c": "retro", + "choice_d": "trailer", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24044 + }, + { + "path": "619272.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "movie", + "choice_b": "soundscape", + "choice_c": "meditative", + "choice_d": "positive", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24045 + }, + { + "path": "1054374.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "love", + "choice_b": "romantic", + "choice_c": "calm", + "choice_d": "groovy", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24046 + }, + { + "path": "1415943.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "dream", + "choice_b": "movie", + "choice_c": "motivational", + "choice_d": "powerful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24047 + }, + { + "path": "1046881.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "cool", + "choice_b": "party", + "choice_c": "inspiring", + "choice_d": "epic", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24048 + }, + { + "path": "1389035.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "sexy", + "choice_b": "emotional", + "choice_c": "nature", + "choice_d": "ballad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24049 + }, + { + "path": "176091.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "drama", + "choice_b": "sexy", + "choice_c": "hopeful", + "choice_d": "travel", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24050 + }, + { + "path": "975527.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "trailer", + "choice_b": "drama", + "choice_c": "sad", + "choice_d": "commercial", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24051 + }, + { + "path": "1116412.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "uplifting", + "choice_b": "dramatic", + "choice_c": "epic", + "choice_d": "romantic", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24052 + }, + { + "path": "1349103.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "dark", + "choice_b": "relaxing", + "choice_c": "sexy", + "choice_d": "action", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24053 + }, + { + "path": "589411.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "summer", + "choice_b": "action", + "choice_c": "energetic", + "choice_d": "cool", + "answer_gt": "action", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24054 + }, + { + "path": "1295658.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "movie", + "choice_b": "christmas", + "choice_c": "energetic", + "choice_d": "commercial", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24055 + }, + { + "path": "1330157.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "film", + "choice_b": "cool", + "choice_c": "soundscape", + "choice_d": "sport", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24056 + }, + { + "path": "1089271.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "epic", + "choice_b": "relaxing", + "choice_c": "slow", + "choice_d": "sad", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24057 + }, + { + "path": "14020.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "relaxing", + "choice_b": "calm", + "choice_c": "melodic", + "choice_d": "energetic", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24058 + }, + { + "path": "1125513.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "soundscape", + "choice_b": "commercial", + "choice_c": "romantic", + "choice_d": "documentary", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24059 + }, + { + "path": "1319325.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "soundscape", + "choice_b": "space", + "choice_c": "travel", + "choice_d": "advertising", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24060 + }, + { + "path": "1370952.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "film", + "choice_b": "meditative", + "choice_c": "action", + "choice_d": "deep", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24061 + }, + { + "path": "1131183.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "energetic", + "choice_b": "nature", + "choice_c": "heavy", + "choice_d": "holiday", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24062 + }, + { + "path": "1117703.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "sad", + "choice_b": "slow", + "choice_c": "deep", + "choice_d": "calm", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24063 + }, + { + "path": "433639.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "meditative", + "choice_b": "soundscape", + "choice_c": "inspiring", + "choice_d": "party", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24064 + }, + { + "path": "1009695.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "retro", + "choice_b": "emotional", + "choice_c": "sport", + "choice_d": "christmas", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24065 + }, + { + "path": "1291569.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "motivational", + "choice_b": "advertising", + "choice_c": "christmas", + "choice_d": "commercial", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24066 + }, + { + "path": "356426.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "sad", + "choice_b": "space", + "choice_c": "christmas", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24067 + }, + { + "path": "756227.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "sexy", + "choice_b": "meditative", + "choice_c": "energetic", + "choice_d": "epic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24068 + }, + { + "path": "1163775.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "retro", + "choice_b": "fun", + "choice_c": "soft", + "choice_d": "christmas", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24069 + }, + { + "path": "53217.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "documentary", + "choice_b": "hopeful", + "choice_c": "energetic", + "choice_d": "dark", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24070 + }, + { + "path": "1056394.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "melodic", + "choice_b": "love", + "choice_c": "dream", + "choice_d": "happy", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24071 + }, + { + "path": "1369436.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "christmas", + "choice_b": "trailer", + "choice_c": "sad", + "choice_d": "commercial", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24072 + }, + { + "path": "865769.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "powerful", + "choice_b": "christmas", + "choice_c": "dramatic", + "choice_d": "action", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24073 + }, + { + "path": "1205197.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "adventure", + "choice_b": "love", + "choice_c": "dramatic", + "choice_d": "romantic", + "answer_gt": "dramatic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24074 + }, + { + "path": "1056391.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "calm", + "choice_b": "meditative", + "choice_c": "dream", + "choice_d": "party", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24075 + }, + { + "path": "883569.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "children", + "choice_b": "epic", + "choice_c": "motivational", + "choice_d": "trailer", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24076 + }, + { + "path": "1349518.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "cool", + "choice_b": "relaxing", + "choice_c": "children", + "choice_d": "dark", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24077 + }, + { + "path": "1083417.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "corporate", + "choice_b": "christmas", + "choice_c": "background", + "choice_d": "game", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24078 + }, + { + "path": "752299.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "powerful", + "choice_b": "sport", + "choice_c": "upbeat", + "choice_d": "background", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24079 + }, + { + "path": "1333717.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "movie", + "choice_b": "retro", + "choice_c": "dream", + "choice_d": "powerful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24080 + }, + { + "path": "1371512.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "summer", + "choice_b": "happy", + "choice_c": "trailer", + "choice_d": "space", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24081 + }, + { + "path": "1354439.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "trailer", + "choice_b": "adventure", + "choice_c": "holiday", + "choice_d": "retro", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24082 + }, + { + "path": "1093614.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "cool", + "choice_b": "drama", + "choice_c": "energetic", + "choice_d": "fast", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24083 + }, + { + "path": "1363499.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "energetic", + "choice_b": "groovy", + "choice_c": "drama", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24084 + }, + { + "path": "1167425.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "love", + "choice_b": "dramatic", + "choice_c": "dark", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24085 + }, + { + "path": "1094477.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "energetic", + "choice_b": "background", + "choice_c": "commercial", + "choice_d": "hopeful", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24086 + }, + { + "path": "146598.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "melodic", + "choice_b": "dream", + "choice_c": "slow", + "choice_d": "uplifting", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24087 + }, + { + "path": "1351058.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "fun", + "choice_b": "upbeat", + "choice_c": "sexy", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24088 + }, + { + "path": "1303934.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "soundscape", + "choice_b": "travel", + "choice_c": "summer", + "choice_d": "deep", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24089 + }, + { + "path": "747437.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "positive", + "choice_b": "soundscape", + "choice_c": "space", + "choice_d": "relaxing", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24090 + }, + { + "path": "1339250.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "dark", + "choice_b": "deep", + "choice_c": "drama", + "choice_d": "sexy", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24091 + }, + { + "path": "1280381.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "hopeful", + "choice_b": "corporate", + "choice_c": "slow", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24092 + }, + { + "path": "1236363.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "positive", + "choice_b": "soft", + "choice_c": "sexy", + "choice_d": "christmas", + "answer_gt": "sexy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24093 + }, + { + "path": "1023271.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "epic", + "choice_b": "advertising", + "choice_c": "energetic", + "choice_d": "documentary", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24094 + }, + { + "path": "398189.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "nature", + "choice_b": "melancholic", + "choice_c": "advertising", + "choice_d": "powerful", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24095 + }, + { + "path": "130696.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "drama", + "choice_b": "meditative", + "choice_c": "slow", + "choice_d": "fun", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24096 + }, + { + "path": "824134.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "holiday", + "choice_b": "happy", + "choice_c": "melodic", + "choice_d": "emotional", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24097 + }, + { + "path": "1397346.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "dramatic", + "choice_b": "emotional", + "choice_c": "happy", + "choice_d": "dark", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24098 + }, + { + "path": "1159861.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "meditative", + "choice_b": "motivational", + "choice_c": "fast", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24099 + }, + { + "path": "87476.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "slow", + "choice_b": "dream", + "choice_c": "uplifting", + "choice_d": "retro", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24100 + }, + { + "path": "1131998.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "dramatic", + "choice_b": "epic", + "choice_c": "nature", + "choice_d": "film", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24101 + }, + { + "path": "91822.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "hopeful", + "choice_b": "advertising", + "choice_c": "commercial", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24102 + }, + { + "path": "1397776.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "relaxing", + "choice_b": "game", + "choice_c": "holiday", + "choice_d": "summer", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24103 + }, + { + "path": "44670.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "energetic", + "choice_b": "adventure", + "choice_c": "slow", + "choice_d": "trailer", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24104 + }, + { + "path": "979236.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "fun", + "choice_b": "epic", + "choice_c": "energetic", + "choice_d": "children", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24105 + }, + { + "path": "887205.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "positive", + "choice_b": "relaxing", + "choice_c": "space", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24106 + }, + { + "path": "173545.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "nature", + "choice_b": "melodic", + "choice_c": "holiday", + "choice_d": "background", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24107 + }, + { + "path": "1348656.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "sexy", + "choice_b": "holiday", + "choice_c": "adventure", + "choice_d": "emotional", + "answer_gt": "sexy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24108 + }, + { + "path": "1280358.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "uplifting", + "choice_b": "calm", + "choice_c": "deep", + "choice_d": "sport", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24109 + }, + { + "path": "1116449.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "children", + "choice_b": "film", + "choice_c": "energetic", + "choice_d": "uplifting", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24110 + }, + { + "path": "1278181.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "film", + "choice_b": "retro", + "choice_c": "dream", + "choice_d": "love", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24111 + }, + { + "path": "1366005.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "hopeful", + "choice_b": "heavy", + "choice_c": "happy", + "choice_d": "background", + "answer_gt": "hopeful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24112 + }, + { + "path": "1420597.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "dark", + "choice_b": "film", + "choice_c": "deep", + "choice_d": "nature", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24113 + }, + { + "path": "1374333.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "powerful", + "choice_b": "upbeat", + "choice_c": "action", + "choice_d": "emotional", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24114 + }, + { + "path": "33830.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "sad", + "choice_b": "uplifting", + "choice_c": "soft", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24115 + }, + { + "path": "1187642.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "commercial", + "choice_b": "sad", + "choice_c": "advertising", + "choice_d": "positive", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24116 + }, + { + "path": "126142.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "hopeful", + "choice_b": "calm", + "choice_c": "soundscape", + "choice_d": "sexy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24117 + }, + { + "path": "1006548.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "soundscape", + "choice_b": "nature", + "choice_c": "relaxing", + "choice_d": "action", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24118 + }, + { + "path": "474785.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "cool", + "choice_b": "dramatic", + "choice_c": "positive", + "choice_d": "sad", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24119 + }, + { + "path": "132216.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "heavy", + "choice_b": "dark", + "choice_c": "dream", + "choice_d": "fast", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24120 + }, + { + "path": "1178112.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "heavy", + "choice_b": "soft", + "choice_c": "slow", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24121 + }, + { + "path": "1412961.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "dream", + "choice_b": "space", + "choice_c": "romantic", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24122 + }, + { + "path": "973747.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "christmas", + "choice_b": "dramatic", + "choice_c": "adventure", + "choice_d": "fast", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24123 + }, + { + "path": "1295654.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "commercial", + "choice_b": "deep", + "choice_c": "christmas", + "choice_d": "party", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24124 + }, + { + "path": "1070643.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "christmas", + "choice_b": "travel", + "choice_c": "groovy", + "choice_d": "powerful", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24125 + }, + { + "path": "1063681.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "holiday", + "choice_b": "deep", + "choice_c": "soundscape", + "choice_d": "sexy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24126 + }, + { + "path": "887203.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "drama", + "choice_b": "melancholic", + "choice_c": "advertising", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24127 + }, + { + "path": "1286511.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "romantic", + "choice_b": "meditative", + "choice_c": "fun", + "choice_d": "sexy", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24128 + }, + { + "path": "1210890.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "ballad", + "choice_b": "game", + "choice_c": "advertising", + "choice_d": "nature", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24129 + }, + { + "path": "921584.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "soft", + "choice_b": "powerful", + "choice_c": "dream", + "choice_d": "epic", + "answer_gt": "soft", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24130 + }, + { + "path": "25222.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "cool", + "choice_b": "soundscape", + "choice_c": "melancholic", + "choice_d": "sport", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24131 + }, + { + "path": "1232469.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "christmas", + "choice_b": "travel", + "choice_c": "documentary", + "choice_d": "dark", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24132 + }, + { + "path": "1381257.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "corporate", + "choice_b": "space", + "choice_c": "heavy", + "choice_d": "advertising", + "answer_gt": "corporate", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24133 + }, + { + "path": "1176716.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "inspiring", + "choice_b": "space", + "choice_c": "advertising", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24134 + }, + { + "path": "146596.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "cool", + "choice_b": "meditative", + "choice_c": "love", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24135 + }, + { + "path": "1276529.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "adventure", + "choice_b": "energetic", + "choice_c": "groovy", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24136 + }, + { + "path": "786205.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "upbeat", + "choice_b": "dream", + "choice_c": "relaxing", + "choice_d": "commercial", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24137 + }, + { + "path": "1072571.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "drama", + "choice_b": "groovy", + "choice_c": "fast", + "choice_d": "heavy", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24138 + }, + { + "path": "784085.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "sport", + "choice_b": "dream", + "choice_c": "hopeful", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24139 + }, + { + "path": "1036107.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "travel", + "choice_b": "epic", + "choice_c": "inspiring", + "choice_d": "movie", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24140 + }, + { + "path": "1353779.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "groovy", + "choice_b": "funny", + "choice_c": "fast", + "choice_d": "emotional", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24141 + }, + { + "path": "747440.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "adventure", + "choice_b": "deep", + "choice_c": "soundscape", + "choice_d": "heavy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24142 + }, + { + "path": "1280924.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "soundscape", + "choice_b": "nature", + "choice_c": "calm", + "choice_d": "game", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24143 + }, + { + "path": "1301595.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "happy", + "choice_b": "film", + "choice_c": "romantic", + "choice_d": "emotional", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24144 + }, + { + "path": "1344193.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "melodic", + "choice_b": "game", + "choice_c": "emotional", + "choice_d": "party", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24145 + }, + { + "path": "146595.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "drama", + "choice_b": "movie", + "choice_c": "sexy", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24146 + }, + { + "path": "217849.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "dream", + "choice_b": "fun", + "choice_c": "melancholic", + "choice_d": "heavy", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24147 + }, + { + "path": "965309.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "romantic", + "choice_b": "space", + "choice_c": "motivational", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24148 + }, + { + "path": "883570.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "calm", + "choice_b": "film", + "choice_c": "epic", + "choice_d": "uplifting", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24149 + }, + { + "path": "258177.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "space", + "choice_b": "travel", + "choice_c": "party", + "choice_d": "melancholic", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24150 + }, + { + "path": "1295659.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "drama", + "choice_b": "christmas", + "choice_c": "motivational", + "choice_d": "fast", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24151 + }, + { + "path": "1351475.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "nature", + "choice_b": "holiday", + "choice_c": "movie", + "choice_d": "retro", + "answer_gt": "holiday", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24152 + }, + { + "path": "1170729.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "deep", + "choice_b": "emotional", + "choice_c": "fast", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24153 + }, + { + "path": "1331766.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "melancholic", + "choice_b": "sexy", + "choice_c": "fun", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24154 + }, + { + "path": "1004121.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "funny", + "choice_b": "film", + "choice_c": "corporate", + "choice_d": "relaxing", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24155 + }, + { + "path": "116237.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "nature", + "choice_b": "travel", + "choice_c": "melodic", + "choice_d": "sad", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24156 + }, + { + "path": "264390.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "space", + "choice_b": "sad", + "choice_c": "deep", + "choice_d": "love", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24157 + }, + { + "path": "1348077.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "heavy", + "choice_b": "sexy", + "choice_c": "adventure", + "choice_d": "inspiring", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24158 + }, + { + "path": "1205195.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "dramatic", + "choice_b": "uplifting", + "choice_c": "trailer", + "choice_d": "upbeat", + "answer_gt": "dramatic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24159 + }, + { + "path": "1066334.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "christmas", + "choice_b": "groovy", + "choice_c": "melodic", + "choice_d": "film", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24160 + }, + { + "path": "1371330.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "cool", + "choice_b": "documentary", + "choice_c": "holiday", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24161 + }, + { + "path": "6606.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "corporate", + "choice_b": "hopeful", + "choice_c": "melodic", + "choice_d": "dark", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24162 + }, + { + "path": "21860.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "christmas", + "choice_b": "heavy", + "choice_c": "sexy", + "choice_d": "happy", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24163 + }, + { + "path": "1348079.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "dream", + "choice_b": "adventure", + "choice_c": "space", + "choice_d": "trailer", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24164 + }, + { + "path": "973753.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "movie", + "choice_b": "fun", + "choice_c": "groovy", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24165 + }, + { + "path": "1167068.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "film", + "choice_b": "drama", + "choice_c": "deep", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24166 + }, + { + "path": "979232.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "fun", + "choice_b": "melancholic", + "choice_c": "funny", + "choice_d": "powerful", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24167 + }, + { + "path": "1165087.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "children", + "choice_b": "film", + "choice_c": "funny", + "choice_d": "sad", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24168 + }, + { + "path": "1410635.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "christmas", + "choice_b": "film", + "choice_c": "happy", + "choice_d": "ballad", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24169 + }, + { + "path": "1347953.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "retro", + "choice_b": "motivational", + "choice_c": "adventure", + "choice_d": "slow", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24170 + }, + { + "path": "1159858.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "relaxing", + "choice_b": "sport", + "choice_c": "dream", + "choice_d": "romantic", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24171 + }, + { + "path": "365770.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "fun", + "choice_b": "film", + "choice_c": "heavy", + "choice_d": "dream", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24172 + }, + { + "path": "1037363.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "fast", + "choice_b": "documentary", + "choice_c": "advertising", + "choice_d": "inspiring", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24173 + }, + { + "path": "1350827.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "children", + "choice_b": "groovy", + "choice_c": "melancholic", + "choice_d": "summer", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24174 + }, + { + "path": "261732.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "dramatic", + "choice_b": "soft", + "choice_c": "melodic", + "choice_d": "sport", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24175 + }, + { + "path": "1400503.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "sexy", + "choice_b": "movie", + "choice_c": "relaxing", + "choice_d": "calm", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24176 + }, + { + "path": "1319506.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "ballad", + "choice_b": "documentary", + "choice_c": "dramatic", + "choice_d": "space", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24177 + }, + { + "path": "1302264.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "christmas", + "choice_b": "deep", + "choice_c": "inspiring", + "choice_d": "sad", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24178 + }, + { + "path": "1053505.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "sexy", + "choice_b": "energetic", + "choice_c": "drama", + "choice_d": "corporate", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24179 + }, + { + "path": "1087538.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "christmas", + "choice_b": "hopeful", + "choice_c": "calm", + "choice_d": "romantic", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24180 + }, + { + "path": "982408.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "melancholic", + "choice_b": "film", + "choice_c": "christmas", + "choice_d": "dream", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24181 + }, + { + "path": "1348976.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "holiday", + "choice_b": "meditative", + "choice_c": "sexy", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24182 + }, + { + "path": "597315.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "summer", + "choice_b": "relaxing", + "choice_c": "drama", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24183 + }, + { + "path": "34696.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "film", + "choice_b": "soundscape", + "choice_c": "melancholic", + "choice_d": "fast", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24184 + }, + { + "path": "967566.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "soundscape", + "choice_b": "happy", + "choice_c": "groovy", + "choice_d": "game", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24185 + }, + { + "path": "414046.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "soundscape", + "choice_b": "inspiring", + "choice_c": "emotional", + "choice_d": "deep", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24186 + }, + { + "path": "1155060.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "energetic", + "choice_b": "film", + "choice_c": "adventure", + "choice_d": "positive", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24187 + }, + { + "path": "1362272.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "powerful", + "choice_b": "heavy", + "choice_c": "epic", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24188 + }, + { + "path": "264329.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "summer", + "choice_b": "retro", + "choice_c": "sad", + "choice_d": "dream", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24189 + }, + { + "path": "189368.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "travel", + "choice_b": "calm", + "choice_c": "action", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24190 + }, + { + "path": "1174090.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "soundscape", + "choice_b": "action", + "choice_c": "game", + "choice_d": "cool", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24191 + }, + { + "path": "1314985.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "soundscape", + "choice_b": "sexy", + "choice_c": "party", + "choice_d": "drama", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24192 + }, + { + "path": "14018.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "space", + "choice_b": "melancholic", + "choice_c": "relaxing", + "choice_d": "travel", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24193 + }, + { + "path": "965306.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "film", + "choice_b": "party", + "choice_c": "summer", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24194 + }, + { + "path": "824624.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "meditative", + "choice_b": "energetic", + "choice_c": "soft", + "choice_d": "romantic", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24195 + }, + { + "path": "1353459.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "dramatic", + "choice_b": "slow", + "choice_c": "action", + "choice_d": "funny", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24196 + }, + { + "path": "1342724.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "cool", + "choice_b": "melodic", + "choice_c": "trailer", + "choice_d": "inspiring", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24197 + }, + { + "path": "1314989.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "soundscape", + "choice_b": "slow", + "choice_c": "trailer", + "choice_d": "inspiring", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24198 + }, + { + "path": "457519.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "ballad", + "choice_b": "inspiring", + "choice_c": "dramatic", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24199 + }, + { + "path": "1037131.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "love", + "choice_b": "fun", + "choice_c": "game", + "choice_d": "hopeful", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24200 + }, + { + "path": "1271205.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "upbeat", + "choice_b": "romantic", + "choice_c": "nature", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24201 + }, + { + "path": "493293.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "soft", + "choice_b": "action", + "choice_c": "children", + "choice_d": "fun", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24202 + }, + { + "path": "1199625.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "uplifting", + "choice_b": "romantic", + "choice_c": "summer", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24203 + }, + { + "path": "1408401.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "nature", + "choice_b": "movie", + "choice_c": "ballad", + "choice_d": "melancholic", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24204 + }, + { + "path": "1037901.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "emotional", + "choice_b": "nature", + "choice_c": "calm", + "choice_d": "background", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24205 + }, + { + "path": "477306.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "soundscape", + "choice_b": "meditative", + "choice_c": "nature", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24206 + }, + { + "path": "1277274.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "party", + "choice_b": "emotional", + "choice_c": "soundscape", + "choice_d": "trailer", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24207 + }, + { + "path": "1330195.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "sexy", + "choice_b": "dream", + "choice_c": "commercial", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24208 + }, + { + "path": "1291568.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "funny", + "choice_b": "epic", + "choice_c": "documentary", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24209 + }, + { + "path": "1033908.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "trailer", + "choice_b": "commercial", + "choice_c": "nature", + "choice_d": "emotional", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24210 + }, + { + "path": "1157350.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "christmas", + "choice_b": "epic", + "choice_c": "calm", + "choice_d": "soundscape", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24211 + }, + { + "path": "549220.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "drama", + "choice_b": "happy", + "choice_c": "hopeful", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24212 + }, + { + "path": "1354151.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "groovy", + "choice_b": "funny", + "choice_c": "movie", + "choice_d": "uplifting", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24213 + }, + { + "path": "1276524.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "soundscape", + "choice_b": "fast", + "choice_c": "nature", + "choice_d": "heavy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24214 + }, + { + "path": "1375902.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "party", + "choice_b": "meditative", + "choice_c": "upbeat", + "choice_d": "happy", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24215 + }, + { + "path": "619270.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "fun", + "choice_b": "sexy", + "choice_c": "soundscape", + "choice_d": "adventure", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24216 + }, + { + "path": "1203415.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "sport", + "choice_b": "ballad", + "choice_c": "corporate", + "choice_d": "travel", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24217 + }, + { + "path": "1033910.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "trailer", + "choice_b": "background", + "choice_c": "groovy", + "choice_d": "emotional", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24218 + }, + { + "path": "1400502.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "inspiring", + "choice_b": "game", + "choice_c": "soft", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24219 + }, + { + "path": "756226.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "dream", + "choice_b": "melancholic", + "choice_c": "space", + "choice_d": "christmas", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24220 + }, + { + "path": "741477.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "calm", + "choice_b": "love", + "choice_c": "upbeat", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24221 + }, + { + "path": "875374.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "soundscape", + "choice_b": "calm", + "choice_c": "christmas", + "choice_d": "emotional", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24222 + }, + { + "path": "1294934.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "meditative", + "choice_b": "soundscape", + "choice_c": "christmas", + "choice_d": "slow", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24223 + }, + { + "path": "1084922.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "melodic", + "choice_b": "cool", + "choice_c": "christmas", + "choice_d": "sad", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24224 + }, + { + "path": "1314983.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "fun", + "choice_b": "dream", + "choice_c": "documentary", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24225 + }, + { + "path": "1053510.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "action", + "choice_b": "ballad", + "choice_c": "film", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24226 + }, + { + "path": "1339597.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "funny", + "choice_b": "sexy", + "choice_c": "energetic", + "choice_d": "sad", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24227 + }, + { + "path": "1155061.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "action", + "choice_b": "energetic", + "choice_c": "dramatic", + "choice_d": "fun", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24228 + }, + { + "path": "945025.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "drama", + "choice_b": "funny", + "choice_c": "emotional", + "choice_d": "commercial", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24229 + }, + { + "path": "1420630.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "powerful", + "choice_b": "holiday", + "choice_c": "corporate", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24230 + }, + { + "path": "1084488.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "epic", + "choice_b": "christmas", + "choice_c": "film", + "choice_d": "nature", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24231 + }, + { + "path": "752315.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "game", + "choice_b": "meditative", + "choice_c": "holiday", + "choice_d": "background", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24232 + }, + { + "path": "1319499.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "groovy", + "choice_b": "travel", + "choice_c": "soundscape", + "choice_d": "background", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24233 + }, + { + "path": "680607.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "holiday", + "choice_b": "soundscape", + "choice_c": "ballad", + "choice_d": "dramatic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24234 + }, + { + "path": "1047959.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "retro", + "choice_b": "dark", + "choice_c": "space", + "choice_d": "romantic", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24235 + }, + { + "path": "1370945.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "melodic", + "choice_b": "game", + "choice_c": "soundscape", + "choice_d": "children", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24236 + }, + { + "path": "1221465.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "cool", + "choice_b": "dark", + "choice_c": "film", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24237 + }, + { + "path": "14028.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "game", + "choice_b": "emotional", + "choice_c": "relaxing", + "choice_d": "calm", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24238 + }, + { + "path": "84494.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "retro", + "choice_b": "happy", + "choice_c": "heavy", + "choice_d": "cool", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24239 + }, + { + "path": "1107282.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "romantic", + "choice_b": "relaxing", + "choice_c": "groovy", + "choice_d": "love", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24240 + }, + { + "path": "1332221.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "energetic", + "choice_b": "travel", + "choice_c": "hopeful", + "choice_d": "adventure", + "answer_gt": "hopeful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24241 + }, + { + "path": "1119884.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "epic", + "choice_b": "meditative", + "choice_c": "positive", + "choice_d": "dark", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24242 + }, + { + "path": "290859.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "ballad", + "choice_b": "game", + "choice_c": "travel", + "choice_d": "slow", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24243 + }, + { + "path": "1314219.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "fast", + "choice_b": "soundscape", + "choice_c": "advertising", + "choice_d": "romantic", + "answer_gt": "fast", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24244 + }, + { + "path": "1286515.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "melodic", + "choice_b": "uplifting", + "choice_c": "romantic", + "choice_d": "slow", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24245 + }, + { + "path": "1420632.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "soundscape", + "choice_b": "film", + "choice_c": "motivational", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24246 + }, + { + "path": "979233.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "travel", + "choice_b": "corporate", + "choice_c": "motivational", + "choice_d": "fun", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24247 + }, + { + "path": "1085878.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "background", + "choice_b": "romantic", + "choice_c": "hopeful", + "choice_d": "ballad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24248 + }, + { + "path": "1281242.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "dark", + "choice_b": "meditative", + "choice_c": "documentary", + "choice_d": "powerful", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24249 + }, + { + "path": "824630.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "meditative", + "choice_b": "trailer", + "choice_c": "soft", + "choice_d": "drama", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24250 + }, + { + "path": "108747.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "upbeat", + "choice_b": "positive", + "choice_c": "calm", + "choice_d": "motivational", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24251 + }, + { + "path": "883571.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "epic", + "choice_b": "nature", + "choice_c": "action", + "choice_d": "relaxing", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24252 + }, + { + "path": "809473.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "happy", + "choice_b": "sexy", + "choice_c": "space", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24253 + }, + { + "path": "1037681.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "adventure", + "choice_b": "travel", + "choice_c": "film", + "choice_d": "drama", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24254 + }, + { + "path": "147386.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "uplifting", + "choice_b": "melancholic", + "choice_c": "soundscape", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24255 + }, + { + "path": "905011.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "soft", + "choice_b": "christmas", + "choice_c": "soundscape", + "choice_d": "sexy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24256 + }, + { + "path": "747441.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "happy", + "choice_b": "soundscape", + "choice_c": "game", + "choice_d": "slow", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24257 + }, + { + "path": "1363486.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "soundscape", + "choice_b": "movie", + "choice_c": "inspiring", + "choice_d": "positive", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24258 + }, + { + "path": "1342005.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "epic", + "choice_b": "dark", + "choice_c": "relaxing", + "choice_d": "action", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24259 + }, + { + "path": "1376463.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "advertising", + "choice_b": "dream", + "choice_c": "love", + "choice_d": "positive", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24260 + }, + { + "path": "277778.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "cool", + "choice_b": "melodic", + "choice_c": "emotional", + "choice_d": "sport", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24261 + }, + { + "path": "1209202.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "game", + "choice_b": "love", + "choice_c": "commercial", + "choice_d": "deep", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24262 + }, + { + "path": "1093616.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "emotional", + "choice_b": "energetic", + "choice_c": "summer", + "choice_d": "party", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24263 + }, + { + "path": "1170734.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "christmas", + "choice_b": "melodic", + "choice_c": "trailer", + "choice_d": "romantic", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24264 + }, + { + "path": "1396996.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "trailer", + "choice_b": "sad", + "choice_c": "relaxing", + "choice_d": "epic", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24265 + }, + { + "path": "1280357.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "calm", + "choice_b": "motivational", + "choice_c": "groovy", + "choice_d": "retro", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24266 + }, + { + "path": "786783.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "summer", + "choice_b": "travel", + "choice_c": "dark", + "choice_d": "inspiring", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24267 + }, + { + "path": "1136837.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "action", + "choice_b": "slow", + "choice_c": "travel", + "choice_d": "upbeat", + "answer_gt": "action", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24268 + }, + { + "path": "1290755.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "film", + "choice_b": "cool", + "choice_c": "documentary", + "choice_d": "powerful", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24269 + }, + { + "path": "1131178.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "melancholic", + "choice_b": "sport", + "choice_c": "nature", + "choice_d": "groovy", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24270 + }, + { + "path": "834109.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "space", + "choice_b": "party", + "choice_c": "deep", + "choice_d": "motivational", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24271 + }, + { + "path": "905012.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "nature", + "choice_b": "summer", + "choice_c": "soundscape", + "choice_d": "romantic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24272 + }, + { + "path": "1352908.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "emotional", + "choice_b": "upbeat", + "choice_c": "ballad", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24273 + }, + { + "path": "1036105.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "inspiring", + "choice_b": "soundscape", + "choice_c": "ballad", + "choice_d": "uplifting", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24274 + }, + { + "path": "1420617.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "deep", + "choice_b": "love", + "choice_c": "soft", + "choice_d": "hopeful", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24275 + }, + { + "path": "1081259.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "christmas", + "choice_b": "sport", + "choice_c": "love", + "choice_d": "drama", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24276 + }, + { + "path": "1286516.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "inspiring", + "choice_b": "romantic", + "choice_c": "sad", + "choice_d": "soundscape", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24277 + }, + { + "path": "1110385.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "space", + "choice_b": "deep", + "choice_c": "documentary", + "choice_d": "happy", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24278 + }, + { + "path": "1380114.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "happy", + "choice_b": "background", + "choice_c": "cool", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24279 + }, + { + "path": "1311817.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "melancholic", + "choice_b": "sad", + "choice_c": "corporate", + "choice_d": "energetic", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24280 + }, + { + "path": "146607.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "calm", + "choice_b": "trailer", + "choice_c": "dream", + "choice_d": "positive", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24281 + }, + { + "path": "40630.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "sport", + "choice_b": "powerful", + "choice_c": "dramatic", + "choice_d": "happy", + "answer_gt": "sport", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24282 + }, + { + "path": "677104.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "retro", + "choice_b": "melodic", + "choice_c": "soundscape", + "choice_d": "emotional", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24283 + }, + { + "path": "978577.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "cool", + "choice_b": "slow", + "choice_c": "christmas", + "choice_d": "retro", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24284 + }, + { + "path": "516734.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "upbeat", + "choice_b": "soundscape", + "choice_c": "positive", + "choice_d": "corporate", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24285 + }, + { + "path": "1415804.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "trailer", + "choice_b": "retro", + "choice_c": "corporate", + "choice_d": "background", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24286 + }, + { + "path": "1349135.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "positive", + "choice_b": "dark", + "choice_c": "movie", + "choice_d": "hopeful", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24287 + }, + { + "path": "1096869.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "sport", + "choice_b": "christmas", + "choice_c": "soundscape", + "choice_d": "action", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24288 + }, + { + "path": "1077915.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "positive", + "choice_b": "space", + "choice_c": "retro", + "choice_d": "game", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24289 + }, + { + "path": "1197390.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "romantic", + "choice_b": "party", + "choice_c": "hopeful", + "choice_d": "retro", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24290 + }, + { + "path": "1054368.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "meditative", + "choice_b": "love", + "choice_c": "heavy", + "choice_d": "background", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24291 + }, + { + "path": "1402458.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "melodic", + "choice_b": "love", + "choice_c": "sad", + "choice_d": "upbeat", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24292 + }, + { + "path": "378614.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "heavy", + "choice_b": "ballad", + "choice_c": "powerful", + "choice_d": "sad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24293 + }, + { + "path": "1102806.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "hopeful", + "choice_b": "nature", + "choice_c": "epic", + "choice_d": "corporate", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24294 + }, + { + "path": "977168.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "christmas", + "choice_b": "upbeat", + "choice_c": "space", + "choice_d": "holiday", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24295 + }, + { + "path": "1291558.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "heavy", + "choice_b": "christmas", + "choice_c": "soft", + "choice_d": "summer", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24296 + }, + { + "path": "1094478.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "emotional", + "choice_b": "uplifting", + "choice_c": "adventure", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24297 + }, + { + "path": "53218.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "calm", + "choice_b": "inspiring", + "choice_c": "background", + "choice_d": "groovy", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24298 + }, + { + "path": "46192.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "relaxing", + "choice_b": "epic", + "choice_c": "retro", + "choice_d": "melodic", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24299 + }, + { + "path": "1415941.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "sport", + "choice_b": "fast", + "choice_c": "happy", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24300 + }, + { + "path": "75223.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "energetic", + "choice_b": "love", + "choice_c": "corporate", + "choice_d": "emotional", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24301 + }, + { + "path": "1264980.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "love", + "choice_b": "documentary", + "choice_c": "melancholic", + "choice_d": "inspiring", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24302 + }, + { + "path": "457516.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "energetic", + "choice_b": "meditative", + "choice_c": "emotional", + "choice_d": "relaxing", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24303 + }, + { + "path": "799915.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "background", + "choice_b": "inspiring", + "choice_c": "emotional", + "choice_d": "trailer", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24304 + }, + { + "path": "189569.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "energetic", + "choice_b": "advertising", + "choice_c": "meditative", + "choice_d": "nature", + "answer_gt": "advertising", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24305 + }, + { + "path": "1157351.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "energetic", + "choice_b": "epic", + "choice_c": "upbeat", + "choice_d": "commercial", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24306 + }, + { + "path": "1276531.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "film", + "choice_b": "calm", + "choice_c": "soundscape", + "choice_d": "trailer", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24307 + }, + { + "path": "1292126.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "christmas", + "choice_b": "soft", + "choice_c": "dramatic", + "choice_d": "motivational", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24308 + }, + { + "path": "976706.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "documentary", + "choice_b": "positive", + "choice_c": "trailer", + "choice_d": "space", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24309 + }, + { + "path": "1292118.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "christmas", + "choice_b": "happy", + "choice_c": "adventure", + "choice_d": "sport", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24310 + }, + { + "path": "1280359.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "calm", + "choice_b": "melodic", + "choice_c": "game", + "choice_d": "dramatic", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24311 + }, + { + "path": "1396987.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "ballad", + "choice_b": "love", + "choice_c": "party", + "choice_d": "fun", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24312 + }, + { + "path": "1401588.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "love", + "choice_b": "summer", + "choice_c": "christmas", + "choice_d": "trailer", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24313 + }, + { + "path": "1119810.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "ballad", + "choice_b": "calm", + "choice_c": "energetic", + "choice_d": "christmas", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24314 + }, + { + "path": "1292128.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "funny", + "choice_b": "commercial", + "choice_c": "fun", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24315 + }, + { + "path": "1167069.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "dark", + "choice_b": "christmas", + "choice_c": "soundscape", + "choice_d": "deep", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24316 + }, + { + "path": "1371513.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "slow", + "choice_b": "children", + "choice_c": "summer", + "choice_d": "deep", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24317 + }, + { + "path": "1107022.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "inspiring", + "choice_b": "retro", + "choice_c": "romantic", + "choice_d": "slow", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24318 + }, + { + "path": "1183239.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "slow", + "choice_b": "groovy", + "choice_c": "dream", + "choice_d": "children", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24319 + }, + { + "path": "23867.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "fun", + "choice_b": "ballad", + "choice_c": "powerful", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24320 + }, + { + "path": "1339442.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "movie", + "choice_b": "emotional", + "choice_c": "epic", + "choice_d": "sport", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24321 + }, + { + "path": "1056387.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "dream", + "choice_b": "funny", + "choice_c": "upbeat", + "choice_d": "action", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24322 + }, + { + "path": "1056382.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "summer", + "choice_b": "dream", + "choice_c": "adventure", + "choice_d": "advertising", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24323 + }, + { + "path": "762193.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "sad", + "choice_b": "soft", + "choice_c": "happy", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24324 + }, + { + "path": "1395177.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "soundscape", + "choice_b": "background", + "choice_c": "game", + "choice_d": "groovy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24325 + }, + { + "path": "1174060.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "travel", + "choice_b": "soundscape", + "choice_c": "dramatic", + "choice_d": "motivational", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24326 + }, + { + "path": "1202953.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "children", + "choice_b": "emotional", + "choice_c": "love", + "choice_d": "romantic", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24327 + }, + { + "path": "611731.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "action", + "choice_b": "documentary", + "choice_c": "drama", + "choice_d": "slow", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24328 + }, + { + "path": "433650.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "soundscape", + "choice_b": "positive", + "choice_c": "christmas", + "choice_d": "melancholic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24329 + }, + { + "path": "1235984.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "soft", + "choice_b": "dark", + "choice_c": "funny", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24330 + }, + { + "path": "1154516.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "epic", + "choice_b": "film", + "choice_c": "emotional", + "choice_d": "slow", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24331 + }, + { + "path": "126093.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "romantic", + "choice_b": "soundscape", + "choice_c": "calm", + "choice_d": "hopeful", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24332 + }, + { + "path": "398192.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "melancholic", + "choice_b": "party", + "choice_c": "corporate", + "choice_d": "slow", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24333 + }, + { + "path": "1350851.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "dream", + "choice_b": "ballad", + "choice_c": "emotional", + "choice_d": "travel", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24334 + }, + { + "path": "1376447.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "dream", + "choice_b": "relaxing", + "choice_c": "sad", + "choice_d": "dark", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24335 + }, + { + "path": "126081.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "nature", + "choice_b": "soundscape", + "choice_c": "documentary", + "choice_d": "happy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24336 + }, + { + "path": "1291559.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "christmas", + "choice_b": "dream", + "choice_c": "movie", + "choice_d": "documentary", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24337 + }, + { + "path": "965318.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "adventure", + "choice_b": "groovy", + "choice_c": "christmas", + "choice_d": "deep", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24338 + }, + { + "path": "1396603.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "trailer", + "choice_b": "travel", + "choice_c": "upbeat", + "choice_d": "movie", + "answer_gt": "upbeat", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24339 + }, + { + "path": "1153509.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "slow", + "choice_b": "happy", + "choice_c": "space", + "choice_d": "relaxing", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24340 + }, + { + "path": "1353353.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "fun", + "choice_b": "dark", + "choice_c": "summer", + "choice_d": "christmas", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24341 + }, + { + "path": "1248623.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "fun", + "choice_b": "background", + "choice_c": "happy", + "choice_d": "space", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24342 + }, + { + "path": "695266.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "christmas", + "choice_b": "hopeful", + "choice_c": "sport", + "choice_d": "holiday", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24343 + }, + { + "path": "982406.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "commercial", + "choice_b": "fun", + "choice_c": "retro", + "choice_d": "ballad", + "answer_gt": "commercial", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24344 + }, + { + "path": "208938.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "trailer", + "choice_b": "fast", + "choice_c": "summer", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24345 + }, + { + "path": "33183.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "meditative", + "choice_b": "fun", + "choice_c": "melodic", + "choice_d": "deep", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24346 + }, + { + "path": "1160602.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "melodic", + "choice_b": "drama", + "choice_c": "epic", + "choice_d": "funny", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24347 + }, + { + "path": "938363.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "inspiring", + "choice_b": "fast", + "choice_c": "romantic", + "choice_d": "adventure", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24348 + }, + { + "path": "1084130.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "hopeful", + "choice_b": "action", + "choice_c": "melodic", + "choice_d": "ballad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24349 + }, + { + "path": "283094.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "positive", + "choice_b": "corporate", + "choice_c": "relaxing", + "choice_d": "melodic", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24350 + }, + { + "path": "1163780.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "energetic", + "choice_b": "film", + "choice_c": "retro", + "choice_d": "dark", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24351 + }, + { + "path": "1043920.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "trailer", + "choice_b": "motivational", + "choice_c": "commercial", + "choice_d": "epic", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24352 + }, + { + "path": "1403730.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "sad", + "choice_b": "funny", + "choice_c": "film", + "choice_d": "adventure", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24353 + }, + { + "path": "1401593.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "happy", + "choice_b": "drama", + "choice_c": "fast", + "choice_d": "holiday", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24354 + }, + { + "path": "1350817.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "children", + "choice_b": "action", + "choice_c": "epic", + "choice_d": "hopeful", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24355 + }, + { + "path": "1131186.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "funny", + "choice_b": "nature", + "choice_c": "upbeat", + "choice_d": "summer", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24356 + }, + { + "path": "1354149.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "fast", + "choice_b": "funny", + "choice_c": "ballad", + "choice_d": "positive", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24357 + }, + { + "path": "1354579.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "travel", + "choice_b": "relaxing", + "choice_c": "inspiring", + "choice_d": "adventure", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24358 + }, + { + "path": "784525.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "trailer", + "choice_b": "relaxing", + "choice_c": "energetic", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24359 + }, + { + "path": "25217.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "film", + "choice_b": "sexy", + "choice_c": "movie", + "choice_d": "emotional", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24360 + }, + { + "path": "1277246.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "documentary", + "choice_b": "soundscape", + "choice_c": "romantic", + "choice_d": "funny", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24361 + }, + { + "path": "81196.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "fast", + "choice_b": "funny", + "choice_c": "party", + "choice_d": "film", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24362 + }, + { + "path": "1070701.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "uplifting", + "choice_b": "christmas", + "choice_c": "children", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24363 + }, + { + "path": "1244465.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "meditative", + "choice_b": "sad", + "choice_c": "dark", + "choice_d": "dream", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24364 + }, + { + "path": "1093635.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "advertising", + "choice_b": "game", + "choice_c": "soundscape", + "choice_d": "action", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24365 + }, + { + "path": "1286492.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "retro", + "choice_b": "space", + "choice_c": "children", + "choice_d": "action", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24366 + }, + { + "path": "104259.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "relaxing", + "choice_b": "inspiring", + "choice_c": "cool", + "choice_d": "game", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24367 + }, + { + "path": "80376.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "emotional", + "choice_b": "holiday", + "choice_c": "love", + "choice_d": "slow", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24368 + }, + { + "path": "1315327.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "hopeful", + "choice_b": "upbeat", + "choice_c": "love", + "choice_d": "summer", + "answer_gt": "hopeful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24369 + }, + { + "path": "426330.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "cool", + "choice_b": "film", + "choice_c": "background", + "choice_d": "christmas", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24370 + }, + { + "path": "1298832.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "soundscape", + "choice_b": "romantic", + "choice_c": "fast", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24371 + }, + { + "path": "1108516.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "uplifting", + "choice_b": "background", + "choice_c": "dark", + "choice_d": "commercial", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24372 + }, + { + "path": "875379.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "trailer", + "choice_b": "meditative", + "choice_c": "summer", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24373 + }, + { + "path": "976713.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "powerful", + "choice_b": "melancholic", + "choice_c": "children", + "choice_d": "trailer", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24374 + }, + { + "path": "73078.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "corporate", + "choice_b": "romantic", + "choice_c": "space", + "choice_d": "children", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24375 + }, + { + "path": "1352139.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "positive", + "choice_b": "space", + "choice_c": "soft", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24376 + }, + { + "path": "1094475.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "happy", + "choice_b": "relaxing", + "choice_c": "action", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24377 + }, + { + "path": "1131185.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "nature", + "choice_b": "film", + "choice_c": "meditative", + "choice_d": "dark", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24378 + }, + { + "path": "1115567.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "motivational", + "choice_b": "love", + "choice_c": "deep", + "choice_d": "summer", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24379 + }, + { + "path": "1416572.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "soundscape", + "choice_b": "dark", + "choice_c": "trailer", + "choice_d": "corporate", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24380 + }, + { + "path": "116244.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "movie", + "choice_b": "travel", + "choice_c": "fast", + "choice_d": "nature", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24381 + }, + { + "path": "1131182.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "ballad", + "choice_b": "nature", + "choice_c": "dark", + "choice_d": "melancholic", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24382 + }, + { + "path": "259327.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "positive", + "choice_b": "sad", + "choice_c": "fast", + "choice_d": "melodic", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24383 + }, + { + "path": "1211623.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "sad", + "choice_b": "uplifting", + "choice_c": "dream", + "choice_d": "happy", + "answer_gt": "uplifting", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24384 + }, + { + "path": "25223.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "romantic", + "choice_b": "dream", + "choice_c": "calm", + "choice_d": "movie", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24385 + }, + { + "path": "1086398.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "motivational", + "choice_b": "soft", + "choice_c": "christmas", + "choice_d": "love", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24386 + }, + { + "path": "1280361.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "powerful", + "choice_b": "background", + "choice_c": "slow", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24387 + }, + { + "path": "1354589.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "christmas", + "choice_b": "adventure", + "choice_c": "love", + "choice_d": "heavy", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24388 + }, + { + "path": "1351057.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "children", + "choice_b": "action", + "choice_c": "ballad", + "choice_d": "love", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24389 + }, + { + "path": "172814.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "melancholic", + "choice_b": "groovy", + "choice_c": "melodic", + "choice_d": "travel", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24390 + }, + { + "path": "815715.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "groovy", + "choice_b": "uplifting", + "choice_c": "advertising", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24391 + }, + { + "path": "76987.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "deep", + "choice_b": "energetic", + "choice_c": "melodic", + "choice_d": "action", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24392 + }, + { + "path": "1107063.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "groovy", + "choice_b": "dark", + "choice_c": "cool", + "choice_d": "drama", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24393 + }, + { + "path": "1347939.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "meditative", + "choice_b": "summer", + "choice_c": "inspiring", + "choice_d": "retro", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24394 + }, + { + "path": "1146198.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "party", + "choice_b": "commercial", + "choice_c": "emotional", + "choice_d": "trailer", + "answer_gt": "commercial", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24395 + }, + { + "path": "1414982.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "soundscape", + "choice_b": "holiday", + "choice_c": "game", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24396 + }, + { + "path": "33177.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "motivational", + "choice_b": "retro", + "choice_c": "commercial", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24397 + }, + { + "path": "344321.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "cool", + "choice_b": "drama", + "choice_c": "corporate", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24398 + }, + { + "path": "257179.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "emotional", + "choice_b": "soundscape", + "choice_c": "children", + "choice_d": "relaxing", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24399 + }, + { + "path": "33821.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "soundscape", + "choice_b": "dramatic", + "choice_c": "children", + "choice_d": "space", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24400 + }, + { + "path": "1117567.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "happy", + "choice_b": "travel", + "choice_c": "dream", + "choice_d": "fast", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24401 + }, + { + "path": "1036106.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "soft", + "choice_b": "soundscape", + "choice_c": "emotional", + "choice_d": "sexy", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24402 + }, + { + "path": "1304424.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "inspiring", + "choice_b": "powerful", + "choice_c": "epic", + "choice_d": "party", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24403 + }, + { + "path": "1352138.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "sad", + "choice_b": "children", + "choice_c": "summer", + "choice_d": "inspiring", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24404 + }, + { + "path": "1193402.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "slow", + "choice_b": "melancholic", + "choice_c": "dream", + "choice_d": "motivational", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24405 + }, + { + "path": "1250838.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "uplifting", + "choice_b": "movie", + "choice_c": "advertising", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24406 + }, + { + "path": "433643.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "meditative", + "choice_b": "soundscape", + "choice_c": "relaxing", + "choice_d": "children", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24407 + }, + { + "path": "1243731.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "hopeful", + "choice_b": "groovy", + "choice_c": "energetic", + "choice_d": "background", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24408 + }, + { + "path": "1353342.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "summer", + "choice_b": "action", + "choice_c": "cool", + "choice_d": "nature", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24409 + }, + { + "path": "1256658.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "drama", + "choice_b": "party", + "choice_c": "corporate", + "choice_d": "melancholic", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24410 + }, + { + "path": "726601.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "fast", + "choice_b": "action", + "choice_c": "uplifting", + "choice_d": "holiday", + "answer_gt": "action", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24411 + }, + { + "path": "1052778.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "groovy", + "choice_b": "trailer", + "choice_c": "energetic", + "choice_d": "relaxing", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24412 + }, + { + "path": "965310.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "upbeat", + "choice_b": "sad", + "choice_c": "christmas", + "choice_d": "inspiring", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24413 + }, + { + "path": "686357.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "adventure", + "choice_b": "epic", + "choice_c": "space", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24414 + }, + { + "path": "756230.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "deep", + "choice_b": "dark", + "choice_c": "funny", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24415 + }, + { + "path": "147393.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "energetic", + "choice_b": "commercial", + "choice_c": "funny", + "choice_d": "upbeat", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24416 + }, + { + "path": "1353341.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "uplifting", + "choice_b": "heavy", + "choice_c": "summer", + "choice_d": "sexy", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24417 + }, + { + "path": "1347952.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "trailer", + "choice_b": "nature", + "choice_c": "retro", + "choice_d": "hopeful", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24418 + }, + { + "path": "1231675.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "background", + "choice_b": "fast", + "choice_c": "positive", + "choice_d": "slow", + "answer_gt": "fast", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24419 + }, + { + "path": "1279077.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "dramatic", + "choice_b": "heavy", + "choice_c": "background", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24420 + }, + { + "path": "984993.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "nature", + "choice_b": "sad", + "choice_c": "love", + "choice_d": "children", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24421 + }, + { + "path": "1402454.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "trailer", + "choice_b": "corporate", + "choice_c": "melodic", + "choice_d": "sexy", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24422 + }, + { + "path": "1167074.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "summer", + "choice_b": "christmas", + "choice_c": "emotional", + "choice_d": "soundscape", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24423 + }, + { + "path": "172817.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "nature", + "choice_b": "commercial", + "choice_c": "motivational", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24424 + }, + { + "path": "1199624.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "love", + "choice_b": "retro", + "choice_c": "commercial", + "choice_d": "groovy", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24425 + }, + { + "path": "126122.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "summer", + "choice_b": "film", + "choice_c": "soundscape", + "choice_d": "inspiring", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24426 + }, + { + "path": "1292130.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "party", + "choice_b": "soft", + "choice_c": "positive", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24427 + }, + { + "path": "376417.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "melodic", + "choice_b": "holiday", + "choice_c": "meditative", + "choice_d": "happy", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24428 + }, + { + "path": "33826.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "sad", + "choice_b": "children", + "choice_c": "game", + "choice_d": "advertising", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24429 + }, + { + "path": "1243718.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "retro", + "choice_b": "travel", + "choice_c": "uplifting", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24430 + }, + { + "path": "1187346.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "retro", + "choice_b": "corporate", + "choice_c": "love", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24431 + }, + { + "path": "824627.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "soundscape", + "choice_b": "epic", + "choice_c": "movie", + "choice_d": "meditative", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24432 + }, + { + "path": "1170740.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "documentary", + "choice_b": "melodic", + "choice_c": "melancholic", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24433 + }, + { + "path": "577044.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "soft", + "choice_b": "ballad", + "choice_c": "love", + "choice_d": "powerful", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24434 + }, + { + "path": "1356470.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "dark", + "choice_b": "corporate", + "choice_c": "powerful", + "choice_d": "motivational", + "answer_gt": "motivational", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24435 + }, + { + "path": "1264981.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "love", + "choice_b": "powerful", + "choice_c": "melancholic", + "choice_d": "adventure", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24436 + }, + { + "path": "1301831.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "dream", + "choice_b": "retro", + "choice_c": "christmas", + "choice_d": "background", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24437 + }, + { + "path": "1010008.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "retro", + "choice_b": "adventure", + "choice_c": "dream", + "choice_d": "advertising", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24438 + }, + { + "path": "1160599.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "melodic", + "choice_b": "upbeat", + "choice_c": "soundscape", + "choice_d": "motivational", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24439 + }, + { + "path": "231611.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "holiday", + "choice_b": "space", + "choice_c": "emotional", + "choice_d": "melancholic", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24440 + }, + { + "path": "375193.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "adventure", + "choice_b": "happy", + "choice_c": "documentary", + "choice_d": "slow", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24441 + }, + { + "path": "1420607.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "deep", + "choice_b": "energetic", + "choice_c": "cool", + "choice_d": "advertising", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24442 + }, + { + "path": "1210887.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "corporate", + "choice_b": "funny", + "choice_c": "ballad", + "choice_d": "party", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24443 + }, + { + "path": "1380713.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "christmas", + "choice_b": "energetic", + "choice_c": "cool", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24444 + }, + { + "path": "896195.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "commercial", + "choice_b": "advertising", + "choice_c": "dream", + "choice_d": "hopeful", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24445 + }, + { + "path": "1356533.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "ballad", + "choice_b": "motivational", + "choice_c": "movie", + "choice_d": "children", + "answer_gt": "motivational", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24446 + }, + { + "path": "1380709.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "love", + "choice_b": "happy", + "choice_c": "game", + "choice_d": "inspiring", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24447 + }, + { + "path": "378617.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "deep", + "choice_b": "soundscape", + "choice_c": "background", + "choice_d": "ballad", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24448 + }, + { + "path": "1199629.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "documentary", + "choice_b": "fun", + "choice_c": "love", + "choice_d": "funny", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24449 + }, + { + "path": "1314218.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "funny", + "choice_b": "advertising", + "choice_c": "heavy", + "choice_d": "love", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24450 + }, + { + "path": "1331835.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "dream", + "choice_b": "travel", + "choice_c": "holiday", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24451 + }, + { + "path": "1250829.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "calm", + "choice_b": "commercial", + "choice_c": "fast", + "choice_d": "corporate", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24452 + }, + { + "path": "1203893.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "slow", + "choice_b": "soft", + "choice_c": "romantic", + "choice_d": "holiday", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24453 + }, + { + "path": "875372.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "soundscape", + "choice_b": "romantic", + "choice_c": "party", + "choice_d": "hopeful", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24454 + }, + { + "path": "965436.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "motivational", + "choice_b": "christmas", + "choice_c": "commercial", + "choice_d": "inspiring", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24455 + }, + { + "path": "516739.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "fun", + "choice_b": "emotional", + "choice_c": "hopeful", + "choice_d": "film", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24456 + }, + { + "path": "1420614.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "deep", + "choice_b": "movie", + "choice_c": "slow", + "choice_d": "inspiring", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24457 + }, + { + "path": "1078362.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "movie", + "choice_b": "christmas", + "choice_c": "action", + "choice_d": "hopeful", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24458 + }, + { + "path": "1289388.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "melancholic", + "choice_b": "dream", + "choice_c": "slow", + "choice_d": "heavy", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24459 + }, + { + "path": "1400507.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "inspiring", + "choice_b": "cool", + "choice_c": "space", + "choice_d": "calm", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24460 + }, + { + "path": "965450.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "funny", + "choice_b": "trailer", + "choice_c": "christmas", + "choice_d": "dark", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24461 + }, + { + "path": "1049834.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "soundscape", + "choice_b": "travel", + "choice_c": "background", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24462 + }, + { + "path": "1163779.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "powerful", + "choice_b": "heavy", + "choice_c": "retro", + "choice_d": "sad", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24463 + }, + { + "path": "1006556.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "travel", + "choice_b": "soundscape", + "choice_c": "funny", + "choice_d": "soft", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24464 + }, + { + "path": "1374322.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "advertising", + "choice_b": "dramatic", + "choice_c": "motivational", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24465 + }, + { + "path": "1078366.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "party", + "choice_b": "trailer", + "choice_c": "christmas", + "choice_d": "background", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24466 + }, + { + "path": "1121562.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "funny", + "choice_b": "meditative", + "choice_c": "uplifting", + "choice_d": "game", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24467 + }, + { + "path": "1415803.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "dream", + "choice_b": "romantic", + "choice_c": "children", + "choice_d": "slow", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24468 + }, + { + "path": "125941.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "drama", + "choice_b": "soundscape", + "choice_c": "groovy", + "choice_d": "background", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24469 + }, + { + "path": "283093.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "sexy", + "choice_b": "dramatic", + "choice_c": "relaxing", + "choice_d": "space", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24470 + }, + { + "path": "1276505.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "adventure", + "choice_b": "relaxing", + "choice_c": "fun", + "choice_d": "melodic", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24471 + }, + { + "path": "768506.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "motivational", + "choice_b": "inspiring", + "choice_c": "epic", + "choice_d": "emotional", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24472 + }, + { + "path": "1187640.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "uplifting", + "choice_b": "game", + "choice_c": "party", + "choice_d": "melancholic", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24473 + }, + { + "path": "875384.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "melancholic", + "choice_b": "soundscape", + "choice_c": "happy", + "choice_d": "fast", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24474 + }, + { + "path": "1309256.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "motivational", + "choice_b": "film", + "choice_c": "fun", + "choice_d": "heavy", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24475 + }, + { + "path": "21859.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "party", + "choice_b": "deep", + "choice_c": "heavy", + "choice_d": "advertising", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24476 + }, + { + "path": "1351476.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "holiday", + "choice_b": "trailer", + "choice_c": "advertising", + "choice_d": "happy", + "answer_gt": "holiday", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24477 + }, + { + "path": "715131.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "soundscape", + "choice_b": "adventure", + "choice_c": "holiday", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24478 + }, + { + "path": "1133480.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "adventure", + "choice_b": "melodic", + "choice_c": "fun", + "choice_d": "heavy", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24479 + }, + { + "path": "1221464.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "love", + "choice_b": "heavy", + "choice_c": "christmas", + "choice_d": "holiday", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24480 + }, + { + "path": "741528.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "summer", + "choice_b": "meditative", + "choice_c": "dream", + "choice_d": "fun", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24481 + }, + { + "path": "1203147.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "space", + "choice_b": "love", + "choice_c": "film", + "choice_d": "commercial", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24482 + }, + { + "path": "24835.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "motivational", + "choice_b": "melodic", + "choice_c": "sexy", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24483 + }, + { + "path": "1235987.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "advertising", + "choice_b": "soundscape", + "choice_c": "melancholic", + "choice_d": "holiday", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24484 + }, + { + "path": "784523.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "deep", + "choice_b": "positive", + "choice_c": "dream", + "choice_d": "christmas", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24485 + }, + { + "path": "128306.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "groovy", + "choice_b": "corporate", + "choice_c": "adventure", + "choice_d": "fast", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24486 + }, + { + "path": "1415940.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "energetic", + "choice_b": "nature", + "choice_c": "deep", + "choice_d": "advertising", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24487 + }, + { + "path": "376419.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "dark", + "choice_b": "romantic", + "choice_c": "uplifting", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24488 + }, + { + "path": "21862.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "holiday", + "choice_b": "deep", + "choice_c": "melodic", + "choice_d": "heavy", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24489 + }, + { + "path": "1280383.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "summer", + "choice_b": "holiday", + "choice_c": "happy", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24490 + }, + { + "path": "1053509.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "melancholic", + "choice_b": "energetic", + "choice_c": "christmas", + "choice_d": "love", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24491 + }, + { + "path": "1032645.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "inspiring", + "choice_b": "motivational", + "choice_c": "corporate", + "choice_d": "heavy", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24492 + }, + { + "path": "815425.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "meditative", + "choice_b": "uplifting", + "choice_c": "melodic", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24493 + }, + { + "path": "1165086.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "powerful", + "choice_b": "fast", + "choice_c": "energetic", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24494 + }, + { + "path": "433642.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "commercial", + "choice_b": "soundscape", + "choice_c": "funny", + "choice_d": "dramatic", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24495 + }, + { + "path": "1354150.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "funny", + "choice_b": "upbeat", + "choice_c": "dramatic", + "choice_d": "soft", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24496 + }, + { + "path": "1351477.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "dream", + "choice_b": "space", + "choice_c": "holiday", + "choice_d": "children", + "answer_gt": "holiday", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24497 + }, + { + "path": "1102807.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "happy", + "choice_b": "uplifting", + "choice_c": "commercial", + "choice_d": "sexy", + "answer_gt": "commercial", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24498 + }, + { + "path": "25241.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "sad", + "choice_b": "action", + "choice_c": "emotional", + "choice_d": "meditative", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24499 + }, + { + "path": "317187.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "dramatic", + "choice_b": "soft", + "choice_c": "groovy", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24500 + }, + { + "path": "1210884.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "melodic", + "choice_b": "film", + "choice_c": "soundscape", + "choice_d": "action", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24501 + }, + { + "path": "1353343.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "summer", + "choice_b": "relaxing", + "choice_c": "powerful", + "choice_d": "positive", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24502 + }, + { + "path": "1349693.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "romantic", + "choice_b": "love", + "choice_c": "positive", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24503 + }, + { + "path": "1163809.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "advertising", + "choice_b": "uplifting", + "choice_c": "groovy", + "choice_d": "happy", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24504 + }, + { + "path": "1354578.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "adventure", + "choice_b": "sad", + "choice_c": "cool", + "choice_d": "space", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24505 + }, + { + "path": "872919.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "energetic", + "choice_b": "dark", + "choice_c": "film", + "choice_d": "documentary", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24506 + }, + { + "path": "1277245.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "dream", + "choice_b": "soundscape", + "choice_c": "romantic", + "choice_d": "happy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24507 + }, + { + "path": "1134758.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "space", + "choice_b": "corporate", + "choice_c": "dramatic", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24508 + }, + { + "path": "1350947.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "corporate", + "choice_b": "children", + "choice_c": "groovy", + "choice_d": "calm", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24509 + }, + { + "path": "990314.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "film", + "choice_b": "adventure", + "choice_c": "deep", + "choice_d": "uplifting", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24510 + }, + { + "path": "1380712.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "emotional", + "choice_b": "sexy", + "choice_c": "heavy", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24511 + }, + { + "path": "1023306.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "documentary", + "choice_b": "fast", + "choice_c": "corporate", + "choice_d": "trailer", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24512 + }, + { + "path": "457517.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "retro", + "choice_b": "deep", + "choice_c": "energetic", + "choice_d": "ballad", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24513 + }, + { + "path": "1006083.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "film", + "choice_b": "game", + "choice_c": "movie", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24514 + }, + { + "path": "1419771.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "slow", + "choice_b": "action", + "choice_c": "background", + "choice_d": "travel", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24515 + }, + { + "path": "1107068.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "inspiring", + "choice_b": "sad", + "choice_c": "dark", + "choice_d": "slow", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24516 + }, + { + "path": "619271.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "action", + "choice_b": "fast", + "choice_c": "soundscape", + "choice_d": "relaxing", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24517 + }, + { + "path": "1351055.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "funny", + "choice_b": "children", + "choice_c": "space", + "choice_d": "motivational", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24518 + }, + { + "path": "1420631.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "deep", + "choice_b": "sport", + "choice_c": "game", + "choice_d": "slow", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24519 + }, + { + "path": "1333715.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "documentary", + "choice_b": "dream", + "choice_c": "adventure", + "choice_d": "travel", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24520 + }, + { + "path": "1136361.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "emotional", + "choice_b": "soundscape", + "choice_c": "ballad", + "choice_d": "happy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24521 + }, + { + "path": "1393535.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "sexy", + "choice_b": "drama", + "choice_c": "game", + "choice_d": "heavy", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24522 + }, + { + "path": "1154298.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "movie", + "choice_b": "space", + "choice_c": "meditative", + "choice_d": "fun", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24523 + }, + { + "path": "1362286.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "calm", + "choice_b": "game", + "choice_c": "advertising", + "choice_d": "summer", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24524 + }, + { + "path": "76976.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "melancholic", + "choice_b": "inspiring", + "choice_c": "epic", + "choice_d": "calm", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24525 + }, + { + "path": "1033909.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "space", + "choice_b": "soft", + "choice_c": "emotional", + "choice_d": "travel", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24526 + }, + { + "path": "976705.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "trailer", + "choice_b": "sexy", + "choice_c": "children", + "choice_d": "travel", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24527 + }, + { + "path": "128299.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "hopeful", + "choice_b": "party", + "choice_c": "slow", + "choice_d": "action", + "answer_gt": "slow", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24528 + }, + { + "path": "786206.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "film", + "choice_b": "dream", + "choice_c": "nature", + "choice_d": "sport", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24529 + }, + { + "path": "976707.mp3", + "question": "What is the emotive tone reflected in this music?", + "choice_a": "trailer", + "choice_b": "sad", + "choice_c": "sexy", + "choice_d": "powerful", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24530 + }, + { + "path": "1108514.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "background", + "choice_b": "epic", + "choice_c": "emotional", + "choice_d": "upbeat", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24531 + }, + { + "path": "1348602.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "dark", + "choice_b": "sexy", + "choice_c": "advertising", + "choice_d": "space", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24532 + }, + { + "path": "701782.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "drama", + "choice_b": "heavy", + "choice_c": "deep", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24533 + }, + { + "path": "1309459.mp3", + "question": "In terms of mood, what does this music suggest?", + "choice_a": "melodic", + "choice_b": "adventure", + "choice_c": "melancholic", + "choice_d": "party", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24534 + }, + { + "path": "1136368.mp3", + "question": "What mood does this composition seem to encapsulate?", + "choice_a": "melancholic", + "choice_b": "inspiring", + "choice_c": "film", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24535 + }, + { + "path": "176084.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "trailer", + "choice_b": "fun", + "choice_c": "advertising", + "choice_d": "emotional", + "answer_gt": "advertising", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24536 + }, + { + "path": "1099748.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "energetic", + "choice_b": "trailer", + "choice_c": "sport", + "choice_d": "upbeat", + "answer_gt": "sport", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24537 + }, + { + "path": "21856.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "heavy", + "choice_b": "melancholic", + "choice_c": "fun", + "choice_d": "travel", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24538 + }, + { + "path": "1318028.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "dream", + "choice_b": "funny", + "choice_c": "happy", + "choice_d": "slow", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24539 + }, + { + "path": "1357699.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "advertising", + "choice_b": "emotional", + "choice_c": "positive", + "choice_d": "cool", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24540 + }, + { + "path": "1107281.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "happy", + "choice_b": "groovy", + "choice_c": "slow", + "choice_d": "dark", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24541 + }, + { + "path": "1225973.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "groovy", + "choice_b": "sad", + "choice_c": "space", + "choice_d": "sexy", + "answer_gt": "sexy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24542 + }, + { + "path": "1070637.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "positive", + "choice_b": "inspiring", + "choice_c": "christmas", + "choice_d": "travel", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24543 + }, + { + "path": "1356318.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "documentary", + "choice_b": "dramatic", + "choice_c": "dream", + "choice_d": "background", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24544 + }, + { + "path": "116246.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "party", + "choice_b": "travel", + "choice_c": "background", + "choice_d": "movie", + "answer_gt": "travel", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24545 + }, + { + "path": "1176656.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "melancholic", + "choice_b": "powerful", + "choice_c": "sport", + "choice_d": "melodic", + "answer_gt": "melancholic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24546 + }, + { + "path": "883574.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "epic", + "choice_b": "upbeat", + "choice_c": "documentary", + "choice_d": "drama", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24547 + }, + { + "path": "114197.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "sad", + "choice_b": "relaxing", + "choice_c": "energetic", + "choice_d": "meditative", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24548 + }, + { + "path": "1222034.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "ballad", + "choice_b": "space", + "choice_c": "meditative", + "choice_d": "fast", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24549 + }, + { + "path": "1309253.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "uplifting", + "choice_b": "upbeat", + "choice_c": "film", + "choice_d": "movie", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24550 + }, + { + "path": "1353456.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "funny", + "choice_b": "holiday", + "choice_c": "heavy", + "choice_d": "party", + "answer_gt": "funny", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24551 + }, + { + "path": "1155056.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "movie", + "choice_b": "energetic", + "choice_c": "christmas", + "choice_d": "space", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24552 + }, + { + "path": "1131686.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "trailer", + "choice_b": "relaxing", + "choice_c": "documentary", + "choice_d": "love", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24553 + }, + { + "path": "1056229.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "nature", + "choice_b": "soundscape", + "choice_c": "hopeful", + "choice_d": "upbeat", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24554 + }, + { + "path": "76977.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "calm", + "choice_b": "retro", + "choice_c": "film", + "choice_d": "melodic", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24555 + }, + { + "path": "123932.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "inspiring", + "choice_b": "happy", + "choice_c": "dark", + "choice_d": "meditative", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24556 + }, + { + "path": "1420595.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "motivational", + "choice_b": "meditative", + "choice_c": "deep", + "choice_d": "love", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24557 + }, + { + "path": "1136360.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "happy", + "choice_b": "calm", + "choice_c": "soundscape", + "choice_d": "emotional", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24558 + }, + { + "path": "76984.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "energetic", + "choice_b": "slow", + "choice_c": "dream", + "choice_d": "summer", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24559 + }, + { + "path": "1402277.mp3", + "question": "How does this music influence the emotional ambiance?", + "choice_a": "soundscape", + "choice_b": "melodic", + "choice_c": "soft", + "choice_d": "dark", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24560 + }, + { + "path": "1032157.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "sport", + "choice_b": "retro", + "choice_c": "soundscape", + "choice_d": "background", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24561 + }, + { + "path": "1362637.mp3", + "question": "What feeling does this score seem to express?", + "choice_a": "party", + "choice_b": "movie", + "choice_c": "sport", + "choice_d": "adventure", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24562 + }, + { + "path": "1100241.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "film", + "choice_b": "action", + "choice_c": "heavy", + "choice_d": "advertising", + "answer_gt": "advertising", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24563 + }, + { + "path": "1420532.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "dream", + "choice_b": "upbeat", + "choice_c": "game", + "choice_d": "calm", + "answer_gt": "game", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24564 + }, + { + "path": "1352148.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "space", + "choice_b": "melodic", + "choice_c": "action", + "choice_d": "children", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24565 + }, + { + "path": "965313.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "christmas", + "choice_b": "children", + "choice_c": "motivational", + "choice_d": "melancholic", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24566 + }, + { + "path": "146600.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "adventure", + "choice_b": "summer", + "choice_c": "happy", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24567 + }, + { + "path": "647257.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "christmas", + "choice_b": "corporate", + "choice_c": "fast", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24568 + }, + { + "path": "1131184.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "positive", + "choice_b": "nature", + "choice_c": "slow", + "choice_d": "sport", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24569 + }, + { + "path": "132205.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "uplifting", + "choice_b": "melodic", + "choice_c": "christmas", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24570 + }, + { + "path": "680610.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "emotional", + "choice_b": "corporate", + "choice_c": "soundscape", + "choice_d": "fast", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24571 + }, + { + "path": "1353346.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "soundscape", + "choice_b": "summer", + "choice_c": "trailer", + "choice_d": "ballad", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24572 + }, + { + "path": "1370942.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "positive", + "choice_b": "fast", + "choice_c": "powerful", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24573 + }, + { + "path": "1149755.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "corporate", + "choice_b": "heavy", + "choice_c": "inspiring", + "choice_d": "dramatic", + "answer_gt": "inspiring", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24574 + }, + { + "path": "44674.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "motivational", + "choice_b": "fast", + "choice_c": "sad", + "choice_d": "slow", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24575 + }, + { + "path": "1280429.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "party", + "choice_b": "sexy", + "choice_c": "christmas", + "choice_d": "trailer", + "answer_gt": "party", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24576 + }, + { + "path": "1119023.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "hopeful", + "choice_b": "love", + "choice_c": "game", + "choice_d": "space", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24577 + }, + { + "path": "17146.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "adventure", + "choice_b": "meditative", + "choice_c": "dark", + "choice_d": "positive", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24578 + }, + { + "path": "1157357.mp3", + "question": "How does this melody play with the listener's emotions?", + "choice_a": "holiday", + "choice_b": "drama", + "choice_c": "trailer", + "choice_d": "love", + "answer_gt": "drama", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24579 + }, + { + "path": "1294942.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "dramatic", + "choice_b": "positive", + "choice_c": "christmas", + "choice_d": "background", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24580 + }, + { + "path": "242953.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "holiday", + "choice_b": "relaxing", + "choice_c": "melodic", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24581 + }, + { + "path": "1348069.mp3", + "question": "What is the emotional color of this music?", + "choice_a": "children", + "choice_b": "dream", + "choice_c": "action", + "choice_d": "emotional", + "answer_gt": "action", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24582 + }, + { + "path": "1349519.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "happy", + "choice_b": "children", + "choice_c": "action", + "choice_d": "advertising", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24583 + }, + { + "path": "620612.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "melodic", + "choice_b": "funny", + "choice_c": "energetic", + "choice_d": "romantic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24584 + }, + { + "path": "620613.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "romantic", + "choice_b": "drama", + "choice_c": "cool", + "choice_d": "commercial", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24585 + }, + { + "path": "875431.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "relaxing", + "choice_b": "soundscape", + "choice_c": "groovy", + "choice_d": "advertising", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24586 + }, + { + "path": "33835.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "deep", + "choice_b": "children", + "choice_c": "party", + "choice_d": "romantic", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24587 + }, + { + "path": "1070639.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "fast", + "choice_b": "slow", + "choice_c": "adventure", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24588 + }, + { + "path": "1136366.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "movie", + "choice_b": "christmas", + "choice_c": "space", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24589 + }, + { + "path": "938364.mp3", + "question": "How would you describe the vibe this music gives off?", + "choice_a": "meditative", + "choice_b": "background", + "choice_c": "epic", + "choice_d": "adventure", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24590 + }, + { + "path": "1167072.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "cool", + "choice_b": "heavy", + "choice_c": "adventure", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24591 + }, + { + "path": "1363498.mp3", + "question": "What is the central mood theme portrayed in this music?", + "choice_a": "relaxing", + "choice_b": "sport", + "choice_c": "soundscape", + "choice_d": "trailer", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24592 + }, + { + "path": "1094516.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "nature", + "choice_b": "melodic", + "choice_c": "powerful", + "choice_d": "holiday", + "answer_gt": "powerful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24593 + }, + { + "path": "78361.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "deep", + "choice_b": "calm", + "choice_c": "nature", + "choice_d": "inspiring", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24594 + }, + { + "path": "1353355.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "groovy", + "choice_b": "upbeat", + "choice_c": "summer", + "choice_d": "commercial", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24595 + }, + { + "path": "887207.mp3", + "question": "What mood does the musical structure of this piece express?", + "choice_a": "energetic", + "choice_b": "cool", + "choice_c": "space", + "choice_d": "deep", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24596 + }, + { + "path": "126062.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "soundscape", + "choice_b": "dark", + "choice_c": "holiday", + "choice_d": "travel", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24597 + }, + { + "path": "1403965.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "love", + "choice_b": "powerful", + "choice_c": "game", + "choice_d": "calm", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24598 + }, + { + "path": "1037362.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "documentary", + "choice_b": "ballad", + "choice_c": "party", + "choice_d": "action", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24599 + }, + { + "path": "1243203.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "summer", + "choice_b": "fun", + "choice_c": "motivational", + "choice_d": "uplifting", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24600 + }, + { + "path": "858541.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "emotional", + "choice_b": "movie", + "choice_c": "ballad", + "choice_d": "romantic", + "answer_gt": "ballad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24601 + }, + { + "path": "46193.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "movie", + "choice_b": "fun", + "choice_c": "relaxing", + "choice_d": "commercial", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24602 + }, + { + "path": "1114849.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "children", + "choice_b": "dream", + "choice_c": "melancholic", + "choice_d": "powerful", + "answer_gt": "powerful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24603 + }, + { + "path": "1187322.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "children", + "choice_b": "calm", + "choice_c": "meditative", + "choice_d": "dramatic", + "answer_gt": "calm", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24604 + }, + { + "path": "1246340.mp3", + "question": "What kind of sentiment does this music invoke?", + "choice_a": "trailer", + "choice_b": "dramatic", + "choice_c": "ballad", + "choice_d": "epic", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24605 + }, + { + "path": "1024416.mp3", + "question": "What atmosphere is created by this musical composition?", + "choice_a": "sport", + "choice_b": "advertising", + "choice_c": "slow", + "choice_d": "sad", + "answer_gt": "sad", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24606 + }, + { + "path": "1408963.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "deep", + "choice_b": "groovy", + "choice_c": "children", + "choice_d": "advertising", + "answer_gt": "groovy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24607 + }, + { + "path": "1291563.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "christmas", + "choice_b": "party", + "choice_c": "retro", + "choice_d": "background", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24608 + }, + { + "path": "1199630.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "meditative", + "choice_b": "uplifting", + "choice_c": "love", + "choice_d": "deep", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24609 + }, + { + "path": "1168499.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "game", + "choice_b": "happy", + "choice_c": "advertising", + "choice_d": "slow", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24610 + }, + { + "path": "1160707.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "space", + "choice_b": "movie", + "choice_c": "retro", + "choice_d": "holiday", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24611 + }, + { + "path": "965319.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "christmas", + "choice_b": "sad", + "choice_c": "inspiring", + "choice_d": "retro", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24612 + }, + { + "path": "1222035.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "meditative", + "choice_b": "uplifting", + "choice_c": "groovy", + "choice_d": "funny", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24613 + }, + { + "path": "1248600.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "nature", + "choice_b": "relaxing", + "choice_c": "sexy", + "choice_d": "funny", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24614 + }, + { + "path": "84496.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "nature", + "choice_b": "romantic", + "choice_c": "happy", + "choice_d": "commercial", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24615 + }, + { + "path": "1354592.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "emotional", + "choice_b": "meditative", + "choice_c": "game", + "choice_d": "adventure", + "answer_gt": "adventure", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24616 + }, + { + "path": "736523.mp3", + "question": "How does this music set a mood for its context?", + "choice_a": "nature", + "choice_b": "sad", + "choice_c": "energetic", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24617 + }, + { + "path": "1301823.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "groovy", + "choice_b": "cool", + "choice_c": "corporate", + "choice_d": "romantic", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24618 + }, + { + "path": "277816.mp3", + "question": "What impression of mood does this music convey?", + "choice_a": "positive", + "choice_b": "emotional", + "choice_c": "groovy", + "choice_d": "space", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24619 + }, + { + "path": "1286509.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "romantic", + "choice_b": "melancholic", + "choice_c": "sexy", + "choice_d": "deep", + "answer_gt": "romantic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24620 + }, + { + "path": "14024.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "advertising", + "choice_b": "relaxing", + "choice_c": "groovy", + "choice_d": "soft", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24621 + }, + { + "path": "1290751.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "documentary", + "choice_b": "background", + "choice_c": "party", + "choice_d": "holiday", + "answer_gt": "documentary", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24622 + }, + { + "path": "715127.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "cool", + "choice_b": "calm", + "choice_c": "space", + "choice_d": "documentary", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24623 + }, + { + "path": "1210871.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "relaxing", + "choice_b": "holiday", + "choice_c": "commercial", + "choice_d": "retro", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24624 + }, + { + "path": "1121565.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "meditative", + "choice_b": "calm", + "choice_c": "dramatic", + "choice_d": "sport", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24625 + }, + { + "path": "1377078.mp3", + "question": "What underlying theme of emotion is present in this music?", + "choice_a": "soundscape", + "choice_b": "positive", + "choice_c": "powerful", + "choice_d": "travel", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24626 + }, + { + "path": "14023.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "soft", + "choice_b": "deep", + "choice_c": "meditative", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24627 + }, + { + "path": "1420589.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "deep", + "choice_b": "nature", + "choice_c": "fast", + "choice_d": "sad", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24628 + }, + { + "path": "1244443.mp3", + "question": "What character of mood does this music embody?", + "choice_a": "dream", + "choice_b": "melodic", + "choice_c": "soundscape", + "choice_d": "children", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24629 + }, + { + "path": "172815.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "sexy", + "choice_b": "fun", + "choice_c": "ballad", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24630 + }, + { + "path": "258148.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "fast", + "choice_b": "retro", + "choice_c": "commercial", + "choice_d": "meditative", + "answer_gt": "meditative", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24631 + }, + { + "path": "1160971.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "nature", + "choice_b": "energetic", + "choice_c": "groovy", + "choice_d": "retro", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24632 + }, + { + "path": "1235967.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "soundscape", + "choice_b": "meditative", + "choice_c": "party", + "choice_d": "summer", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24633 + }, + { + "path": "1133636.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "motivational", + "choice_b": "summer", + "choice_c": "retro", + "choice_d": "energetic", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24634 + }, + { + "path": "1314848.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "dramatic", + "choice_b": "cool", + "choice_c": "space", + "choice_d": "retro", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24635 + }, + { + "path": "1274541.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "children", + "choice_b": "soundscape", + "choice_c": "sad", + "choice_d": "deep", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24636 + }, + { + "path": "1402278.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "soundscape", + "choice_b": "fun", + "choice_c": "calm", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24637 + }, + { + "path": "1306024.mp3", + "question": "What mood does this music set for its audience?", + "choice_a": "background", + "choice_b": "drama", + "choice_c": "hopeful", + "choice_d": "slow", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24638 + }, + { + "path": "1079251.mp3", + "question": "What emotional motif does this music represent?", + "choice_a": "trailer", + "choice_b": "soundscape", + "choice_c": "party", + "choice_d": "inspiring", + "answer_gt": "trailer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24639 + }, + { + "path": "126042.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "advertising", + "choice_b": "soundscape", + "choice_c": "groovy", + "choice_d": "heavy", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24640 + }, + { + "path": "1243729.mp3", + "question": "What is the emotional tone of this melody?", + "choice_a": "soft", + "choice_b": "upbeat", + "choice_c": "relaxing", + "choice_d": "emotional", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24641 + }, + { + "path": "1119808.mp3", + "question": "What vibe is resonated by this piece of music?", + "choice_a": "advertising", + "choice_b": "relaxing", + "choice_c": "happy", + "choice_d": "nature", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24642 + }, + { + "path": "1094522.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "children", + "choice_b": "background", + "choice_c": "uplifting", + "choice_d": "summer", + "answer_gt": "background", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24643 + }, + { + "path": "1294940.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "funny", + "choice_b": "movie", + "choice_c": "christmas", + "choice_d": "sport", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24644 + }, + { + "path": "33179.mp3", + "question": "What is the mood resonance of this musical selection?", + "choice_a": "sexy", + "choice_b": "melodic", + "choice_c": "summer", + "choice_d": "game", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24645 + }, + { + "path": "1170730.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "christmas", + "choice_b": "energetic", + "choice_c": "movie", + "choice_d": "love", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24646 + }, + { + "path": "159663.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "emotional", + "choice_b": "action", + "choice_c": "film", + "choice_d": "fast", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24647 + }, + { + "path": "1163778.mp3", + "question": "How does this music affect the listener's mood?", + "choice_a": "retro", + "choice_b": "uplifting", + "choice_c": "soft", + "choice_d": "fun", + "answer_gt": "retro", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24648 + }, + { + "path": "1131176.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "fast", + "choice_b": "children", + "choice_c": "nature", + "choice_d": "relaxing", + "answer_gt": "nature", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24649 + }, + { + "path": "16101.mp3", + "question": "What is the predominant mood signature of this music?", + "choice_a": "soundscape", + "choice_b": "heavy", + "choice_c": "commercial", + "choice_d": "emotional", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24650 + }, + { + "path": "1221449.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "love", + "choice_b": "fast", + "choice_c": "heavy", + "choice_d": "children", + "answer_gt": "love", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24651 + }, + { + "path": "86742.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "trailer", + "choice_b": "slow", + "choice_c": "film", + "choice_d": "holiday", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24652 + }, + { + "path": "1275302.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "documentary", + "choice_b": "space", + "choice_c": "soundscape", + "choice_d": "positive", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24653 + }, + { + "path": "1420618.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "sport", + "choice_b": "heavy", + "choice_c": "slow", + "choice_d": "deep", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24654 + }, + { + "path": "1339600.mp3", + "question": "How does this musical piece serve as a mood enhancer?", + "choice_a": "dark", + "choice_b": "sad", + "choice_c": "sexy", + "choice_d": "positive", + "answer_gt": "sexy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24655 + }, + { + "path": "1168480.mp3", + "question": "What is the affective theme of this composition?", + "choice_a": "corporate", + "choice_b": "melodic", + "choice_c": "commercial", + "choice_d": "christmas", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24656 + }, + { + "path": "1089269.mp3", + "question": "What feeling does the harmony and melody of this music portray?", + "choice_a": "hopeful", + "choice_b": "epic", + "choice_c": "commercial", + "choice_d": "meditative", + "answer_gt": "epic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24657 + }, + { + "path": "1111209.mp3", + "question": "What kind of emotional response does this tune elicit?", + "choice_a": "ballad", + "choice_b": "christmas", + "choice_c": "energetic", + "choice_d": "advertising", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24658 + }, + { + "path": "1348659.mp3", + "question": "What emotive qualities does this music project?", + "choice_a": "fun", + "choice_b": "soft", + "choice_c": "children", + "choice_d": "sad", + "answer_gt": "fun", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24659 + }, + { + "path": "33832.mp3", + "question": "How does this symphony speak to the mood of the audience?", + "choice_a": "retro", + "choice_b": "melodic", + "choice_c": "children", + "choice_d": "christmas", + "answer_gt": "children", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24660 + }, + { + "path": "128247.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "film", + "choice_b": "dream", + "choice_c": "motivational", + "choice_d": "party", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24661 + }, + { + "path": "858540.mp3", + "question": "What kind of emotional atmosphere does this piece generate?", + "choice_a": "emotional", + "choice_b": "happy", + "choice_c": "advertising", + "choice_d": "party", + "answer_gt": "emotional", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24662 + }, + { + "path": "620998.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "love", + "choice_b": "sad", + "choice_c": "film", + "choice_d": "relaxing", + "answer_gt": "relaxing", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24663 + }, + { + "path": "126120.mp3", + "question": "In what emotional direction does this music take the listener?", + "choice_a": "dramatic", + "choice_b": "soundscape", + "choice_c": "drama", + "choice_d": "commercial", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24664 + }, + { + "path": "1056383.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "film", + "choice_b": "meditative", + "choice_c": "retro", + "choice_d": "dream", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24665 + }, + { + "path": "223134.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "christmas", + "choice_b": "cool", + "choice_c": "dramatic", + "choice_d": "soundscape", + "answer_gt": "cool", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24666 + }, + { + "path": "1132936.mp3", + "question": "What is the prevailing mood theme of this music?", + "choice_a": "inspiring", + "choice_b": "melodic", + "choice_c": "powerful", + "choice_d": "upbeat", + "answer_gt": "powerful", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24667 + }, + { + "path": "1393148.mp3", + "question": "Is there a distinct mood that this music captures?", + "choice_a": "cool", + "choice_b": "christmas", + "choice_c": "love", + "choice_d": "advertising", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24668 + }, + { + "path": "867445.mp3", + "question": "What is the expressive mood tone of this soundtrack?", + "choice_a": "sad", + "choice_b": "corporate", + "choice_c": "dream", + "choice_d": "documentary", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24669 + }, + { + "path": "1056398.mp3", + "question": "How does this music shape the emotional climate?", + "choice_a": "trailer", + "choice_b": "dream", + "choice_c": "background", + "choice_d": "soft", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24670 + }, + { + "path": "1236752.mp3", + "question": "What mood does this musical arrangement translate to?", + "choice_a": "corporate", + "choice_b": "dream", + "choice_c": "upbeat", + "choice_d": "soundscape", + "answer_gt": "dream", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24671 + }, + { + "path": "1327464.mp3", + "question": "What is the mood theme woven into this musical piece?", + "choice_a": "children", + "choice_b": "deep", + "choice_c": "dramatic", + "choice_d": "love", + "answer_gt": "deep", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24672 + }, + { + "path": "16092.mp3", + "question": "How does this piece articulate a mood experience?", + "choice_a": "relaxing", + "choice_b": "heavy", + "choice_c": "adventure", + "choice_d": "drama", + "answer_gt": "heavy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24673 + }, + { + "path": "165311.mp3", + "question": "What is the mood narrative of this music?", + "choice_a": "melodic", + "choice_b": "love", + "choice_c": "documentary", + "choice_d": "christmas", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24674 + }, + { + "path": "1128658.mp3", + "question": "What is the sentiment echoed by this music?", + "choice_a": "fast", + "choice_b": "space", + "choice_c": "drama", + "choice_d": "heavy", + "answer_gt": "space", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24675 + }, + { + "path": "1393143.mp3", + "question": "What emotions does this piece of music evoke?", + "choice_a": "fun", + "choice_b": "christmas", + "choice_c": "fast", + "choice_d": "relaxing", + "answer_gt": "christmas", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24676 + }, + { + "path": "1136367.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "fun", + "choice_b": "powerful", + "choice_c": "summer", + "choice_d": "soundscape", + "answer_gt": "soundscape", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24677 + }, + { + "path": "683572.mp3", + "question": "Which mood theme does this soundtrack primarily convey?", + "choice_a": "drama", + "choice_b": "christmas", + "choice_c": "energetic", + "choice_d": "movie", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24678 + }, + { + "path": "261827.mp3", + "question": "How does this music reflect a particular mood?", + "choice_a": "action", + "choice_b": "film", + "choice_c": "energetic", + "choice_d": "melodic", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24679 + }, + { + "path": "1281248.mp3", + "question": "How does this music contribute to the mood of the moment?", + "choice_a": "dark", + "choice_b": "fun", + "choice_c": "holiday", + "choice_d": "melodic", + "answer_gt": "dark", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24680 + }, + { + "path": "588703.mp3", + "question": "What emotive character does this tune portray?", + "choice_a": "melodic", + "choice_b": "ballad", + "choice_c": "melancholic", + "choice_d": "fast", + "answer_gt": "melodic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24681 + }, + { + "path": "1353349.mp3", + "question": "What mood does the orchestration of this music imply?", + "choice_a": "retro", + "choice_b": "commercial", + "choice_c": "summer", + "choice_d": "fun", + "answer_gt": "summer", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24682 + }, + { + "path": "1232637.mp3", + "question": "What is the emotional essence of this musical work?", + "choice_a": "positive", + "choice_b": "happy", + "choice_c": "calm", + "choice_d": "nature", + "answer_gt": "happy", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24683 + }, + { + "path": "1393011.mp3", + "question": "What emotional landscape does this music paint?", + "choice_a": "motivational", + "choice_b": "dark", + "choice_c": "energetic", + "choice_d": "children", + "answer_gt": "energetic", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24684 + }, + { + "path": "1041289.mp3", + "question": "What feeling does the tempo and rhythm of this music suggest?", + "choice_a": "deep", + "choice_b": "heavy", + "choice_c": "melancholic", + "choice_d": "film", + "answer_gt": "film", + "task_name": "Music_Mood_Recognition", + "dataset_name": "MTJ-Jamendo", + "uniq_id": 24685 + } +]